Patents by Inventor Kshitij A. Doshi

Kshitij A. Doshi has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20220247635
    Abstract: Methods, apparatus, systems and articles of manufacture are disclosed to control processing of telemetry data at an edge platform. An example apparatus includes an orchestrator interface to, responsive to an amount of resources allocated to an orchestrator to orchestrate a workload at the edge platform meeting a first threshold, transmit telemetry data associated with the orchestrator to a computer to obtain a first orchestration result at a first granularity; a resource management controller to determine a second orchestration result at a second granularity to orchestrate the workload at the edge platform, the second granularity finer than the first granularity; and a scheduler to schedule a workload assigned to the edge platform based on the second orchestration result.
    Type: Application
    Filed: October 8, 2021
    Publication date: August 4, 2022
    Inventors: Francesc Guim Bernat, Kshitij Doshi, Ned Smith, Thijs Metsch
  • Patent number: 11402235
    Abstract: A system and method for managing sensors including determining health operation states of the sensors correlative with sensor accuracy, classifying the sensors by their respective health operation state, and teaming two sensors each having a health operation state that is intermediate to give a team having a health operation state that is healthy. The sampling frequency of the sensors to determine sensor accuracy may be dynamic.
    Type: Grant
    Filed: January 3, 2017
    Date of Patent: August 2, 2022
    Assignee: Intel Corporation
    Inventors: Kshitij A. Doshi, Tao Zhong, Gang Deng, Zhongyan Lu
  • Publication number: 20220237033
    Abstract: Technologies for migrating data between edge accelerators hosted on different edge locations include a device hosted on a present edge location. The device includes one or more processors to: receive a workload from a requesting device, determine one or more accelerator devices hosted on the present edge location to perform the workload, and transmit the workload to the one or more accelerator devices to process the workload. The one or more processor is further to determine whether to perform data migration from the one or more accelerator devices to one or more different edge accelerator devices hosted on a different edge location, and send, in response to a determination to perform the data migration, a request to the one or more accelerator devices on the present edge location for transformed workload data to be processed by the one or more different edge accelerator devices.
    Type: Application
    Filed: February 7, 2022
    Publication date: July 28, 2022
    Inventors: Evan Custodio, Francesc Guim Bernat, Suraj Prabhakaran, Trevor Cooper, Ned M. Smith, Kshitij Doshi, Petar Torre
  • Publication number: 20220222118
    Abstract: Methods, apparatus, and systems for adaptive collaborative memory with the assistance of programmable networking devices. Under one example, the programmable networking device is a switch that is deployed in a system or cluster of servers comprising a plurality of nodes. The switch selects one or more nodes to be remote memory server nodes and allocate one or more portions of memory on those nodes to be used as remote memory for one or more remote memory client nodes. The switch receives memory access request messages originating from remote memory client nodes containing indicia identifying memory to be accessed, determines which remote memory server node is to be used for servicing a given memory access request, and sends a memory access request message containing indicia identifying memory to be accessed to the remote memory server node that is determined. The switch also facilitates return of messages containing remote memory access responses to the client nodes.
    Type: Application
    Filed: March 31, 2022
    Publication date: July 14, 2022
    Inventors: Ren WANG, Christian MACIOCCO, Yipeng WANG, Kshitij A. DOSHI, Vesh Raj SHARMA BANJADE, Satish C. JHA, S M Iftekharul ALAM, Srikathyayani SRIKANTESWARA, Alexander BACHMUTSKY
  • Patent number: 11388272
    Abstract: Technologies for network packet processing between cloud and telecommunications networks includes a network computing device which includes two application layer packet translators (ALPTs). The first ALPT is configured to receive a network packet from a computing device in a telecommunications network, identify a virtual network function (VNF) instance, and perform an application layer encapsulation of at least a portion of data of the received network packet as a parameter of a remote procedure call (RPC) associated with the identified VNF instance. The first ALPT is additionally configured to invoke the identified VNF instance using an API call corresponding to the RPC that includes the RPC parameter and the VNF instance is configured to transmit an RPC call response to the second ALPT. The second ALPT is configured to generate a new network packet as a function of the RPC call response and transmit the new network packet to another computing device in a cloud network.
    Type: Grant
    Filed: March 30, 2018
    Date of Patent: July 12, 2022
    Assignee: Intel Corporation
    Inventors: Vadim Sukhomlinov, Kshitij Doshi, Areg Melik-Adamyan
  • Patent number: 11379264
    Abstract: Some examples provide for uninterruptible power supply form (UPS) resources and non-UPS resources to be offered in a composite node for customers to use. For a workload run on the composite node, monitoring of non-UPS resource power availability, resource temperature, and/or cooling facilities can take place. In the event, a non-UPS resource experiences a power outage or reduction in available power, temperature that is at or above a threshold level, and/or cooling facility outage, monitoring of performance of a workload executing on the non-UPS resource can take place. If the performance is acceptable and the power available to the non-UPS resource exceeds a threshold level, the supplied power can be reduced. If the performance experiences excessive levels of errors or slows unacceptably, the workload can be migrated to another non-UPS or UPS compliant resource.
    Type: Grant
    Filed: April 15, 2019
    Date of Patent: July 5, 2022
    Assignee: Intel Corporation
    Inventors: Francesc Guim Bernat, Felipe Pastor Beneyto, Kshitij A. Doshi, Timothy Verrall, Suraj Prabhakaran
  • Publication number: 20220204542
    Abstract: In an example, there is disclosed a host-fabric interface (HFI), including: an interconnect interface to communicatively couple the HFI to an interconnect; a network interface to communicatively couple the HFI to a network; network interface logic to provide communication between the interconnect and the network; a coprocessor configured to provide an offloaded function for the network; a memory; and a caching agent configured to: designate a region of the memory as a shared memory between the HFI and a core communicatively coupled to the HFI via the interconnect; receive a memory operation directed to the shared memory; and issue a memory instruction to the memory according to the memory operation.
    Type: Application
    Filed: October 4, 2021
    Publication date: June 30, 2022
    Applicant: Intel Corporation
    Inventors: Francesc Guim Bernat, Daniel Rivas Barragan, Kshitij A. Doshi, Mark A. Schmisseur
  • Publication number: 20220209971
    Abstract: Methods, apparatus, systems, and articles of manufacture are disclosed to aggregate telemetry data in an edge environment. An example apparatus includes at least one processor, and memory including instructions that, when executed, cause the at least one processor to at least generate a composition for an edge service in the edge environment, the composition representative of a first interface to obtain the telemetry data, the telemetry data associated with resources of the edge service and including a performance metric, generate a resource object based on the performance metric, generate a telemetry object based on the performance metric, and generate a telemetry executable based on the composition, the composition including at least one of the resource object or the telemetry object, the telemetry executable to generate the telemetry data in response to the edge service executing a computing task distributed to the edge service based on the telemetry data.
    Type: Application
    Filed: January 4, 2022
    Publication date: June 30, 2022
    Inventors: Kshitij Doshi, Francesc Guim Bernat, Timothy Verrall, Ned Smith, Rajesh Gadiyar
  • Publication number: 20220198071
    Abstract: Methods, apparatus, systems, and articles of manufacture are disclosed for secure execution on smart network interface cards. An example non-transitory computer readable medium comprising instructions that, when executed, cause a processor to at least: obtain instructions for execution, decode the instructions, execute the decoded instructions, determine an execution frequency, and, when the execution frequency meets a threshold, provide an alert of the execution frequency.
    Type: Application
    Filed: December 23, 2021
    Publication date: June 23, 2022
    Inventors: Kshitij Doshi, Francesc Guim Bernat
  • Publication number: 20220200872
    Abstract: Embodiments may be generally directed to techniques to cause communication of a registration request between a first end-point and a second end-point of an end-to-end path, the registration request to establish resource load monitoring for one or more resources of the end-to-end path, receive one or more acknowledgements indicating resource loads for each of the one or more resources of the end-to-end path, at least one of the acknowledgements to indicate a resource of the one or more resources is not meeting a threshold requirement for the end-to-end path, and perform an action for communication traffic utilizing the one or more resources based on the acknowledgement.
    Type: Application
    Filed: March 9, 2022
    Publication date: June 23, 2022
    Applicant: INTEL CORPORATION
    Inventors: FRANCESC GUIM BERNAT, KSHITIJ A. DOSHI, DANIEL RIVAS BARRAGAN, MARK A. SCHMISSEUR, STEEN LARSEN
  • Patent number: 11354240
    Abstract: The present disclosure is directed to systems and methods that include cache operation storage circuitry that selectively enables/disables the Cache Line Flush (CLFLUSH) operation. The cache operation storage circuitry may also selectively replace the CLFLUSH operation with one or more replacement operations that provide similar functionality but beneficially and advantageously prevent an attacker from placing processor cache circuitry in a known state during a timing-based, side channel attack such as Spectre or Meltdown. The cache operation storage circuitry includes model specific registers (MSRs) that contain information used to determine whether to enable/disable CLFLUSH functionality. The cache operation storage circuitry may include model specific registers (MSRs) that contain information used to select appropriate replacement operations such as Cache Line Demote (CLDEMOTE) and/or Cache Line Write Back (CLWB) to selectively replace CLFLUSH operations.
    Type: Grant
    Filed: June 22, 2020
    Date of Patent: June 7, 2022
    Assignee: Intel Corporation
    Inventors: Vadim Sukhomlinov, Kshitij Doshi
  • Patent number: 11354127
    Abstract: A computing system includes a memory controller having a plurality of bypass parameters set by a software program, a thresholds matrix to store threshold values selectable by the plurality of bypass parameters, and a bypass function to determine whether a first cache line is to be displaced with a second cache line in a first memory or the first cache line remains in the first memory and the second cache line is to be accessed by at least one of a processor core and the cache from a second memory.
    Type: Grant
    Filed: July 13, 2020
    Date of Patent: June 7, 2022
    Assignee: INTEL CORPORATION
    Inventors: Harshad S. Sane, Anup Mohan, Kshitij A. Doshi, Mark A. Schmisseur
  • Publication number: 20220171648
    Abstract: In one embodiment, a computing device comprises memory circuitry and processing circuitry. The memory circuitry is to store a plurality of container images, comprising: a first container image comprising a first set of applications; and a second container image comprising a virtual machine, a guest operating system, and a second set of applications. The processing circuitry is to: instantiate a plurality of containers on a host operating system, wherein the plurality of containers comprises a first container and a second container; execute the first set of applications in the first container, wherein the first set of applications is to be executed on the host operating system; and execute the virtual machine in the second container, wherein the guest operating system is to be executed on the virtual machine and the second set of applications is to be executed on the guest operating system.
    Type: Application
    Filed: May 11, 2020
    Publication date: June 2, 2022
    Applicant: Intel Corporation
    Inventors: Bryan J. Rodriguez, Kshitij A. Doshi, Ned M. Smith, Michael G. Millsap
  • Patent number: 11345342
    Abstract: An apparatus comprising a memory to store an observed trajectory of a pedestrian, the observed trajectory comprising a plurality of observed locations of the pedestrian over a first plurality of timesteps; and a processor to generate a predicted trajectory of the pedestrian, the predicted trajectory comprising a plurality of predicted locations of the pedestrian over the first plurality of timesteps and over a second plurality of timesteps occurring after the first plurality of timesteps; determine a likelihood of the predicted trajectory based on a comparison of the plurality of predicted locations of the pedestrian over the first plurality of timesteps and the plurality of observed locations of the pedestrian over the first plurality of timesteps; and responsive to the determined likelihood of the predicted trajectory, provide information associated with the predicted trajectory to a vehicle to warn the vehicle of a potential collision with the pedestrian.
    Type: Grant
    Filed: September 27, 2019
    Date of Patent: May 31, 2022
    Assignee: Intel Corporation
    Inventors: David Gomez Gutierrez, Javier Felip Leon, Kshitij A. Doshi, Leobardo E. Campos Macias, Nilesh Amar Ahuja, Omesh Tickoo
  • Patent number: 11334382
    Abstract: Technologies for batching requests in an edge infrastructure include a compute device including circuitry configured to obtain a request for an operation to be performed at an edge location. The circuitry is also configured to determine, as a function of a parameter of the obtained request, a batch that the obtained request is to be assigned to. The batch includes a one or more requests for operations to be performed at an edge location. The circuitry is also configured to assign the batch to a cloudlet at an edge location. The cloudlet includes a set of resources usable to execute the operations requested in the batch.
    Type: Grant
    Filed: September 6, 2019
    Date of Patent: May 17, 2022
    Assignee: Intel Corporation
    Inventors: Francesc Guim Bernat, Kshitij Doshi, Suraj Prabhakaran, Ned M. Smith
  • Publication number: 20220147395
    Abstract: There is disclosed an example of an artificial intelligence (AI) system, including: a first hardware platform; a fabric interface configured to communicatively couple the first hardware platform to a second hardware platform; a processor hosted on the first hardware platform and programmed to operate on an AI problem; and a first training accelerator, including: an accelerator hardware; a platform inter-chip link (ICL) configured to communicatively couple the first training accelerator to a second training accelerator on the first hardware platform without aid of the processor; a fabric ICL to communicatively couple the first training accelerator to a third training accelerator on a second hardware platform without aid of the processor; and a system decoder configured to operate the fabric ICL and platform ICL to share data of the accelerator hardware between the first training accelerator and second and third training accelerators without aid of the processor.
    Type: Application
    Filed: January 25, 2022
    Publication date: May 12, 2022
    Applicant: Intel Corporation
    Inventors: Francesc Guim Bernat, Da-Ming Chiang, Kshitij A. Doshi, Suraj Prabhakaran, Mark A. Schmisseur
  • Patent number: 11329898
    Abstract: Embodiments may be generally directed to techniques to cause communication of a registration request between a first end-point and a second end-point of an end-to-end path, the registration request to establish resource load monitoring for one or more resources of the end-to-end path, receive one or more acknowledgements indicating resource loads for each of the one or more resources of the end-to-end path, at least one of the acknowledgements to indicate a resource of the one or more resources is not meeting a threshold requirement for the end-to-end path, and perform an action for communication traffic utilizing the one or more resources based on the acknowledgement.
    Type: Grant
    Filed: March 16, 2021
    Date of Patent: May 10, 2022
    Assignee: INTEL CORPORATION
    Inventors: Francesc Guim Bernat, Kshitij A. Doshi, Daniel Rivas Barragan, Mark A. Schmisseur, Steen Larsen
  • Publication number: 20220138156
    Abstract: Methods, apparatus, systems, and articles of manufacture providing a tiered elastic cloud storage to increase data resiliency are disclosed.
    Type: Application
    Filed: November 8, 2021
    Publication date: May 5, 2022
    Inventors: Ren Wang, Christian Maciocco, Kshitij Doshi, Francesc Guim Bernat, Ned Smith, Satish Jha, Vesh Raj Sharma Banjade, S M Iftekharul Alam
  • Publication number: 20220121470
    Abstract: In one embodiment, metadata associated with deployment of a container within an orchestration environment includes information indicating security preferences for deployment of the container within the orchestration environment, information indicating a level of communications between the container and other containers, and/or information indicating effects of execution of the container with respect to other containers. The metadata is used to select a particular node of a plurality of nodes within the orchestration environment on which to deploy the container based on the metadata.
    Type: Application
    Filed: December 23, 2021
    Publication date: April 21, 2022
    Applicant: Intel Corporation
    Inventors: Paritosh Saxena, Anjo Lucas Vahldiek-Oberwagner, Mona Vij, Kshitij A. Doshi, Carlos H. Morales, Clair Bowman, Marcela S. Melara, Michael Steiner
  • Publication number: 20220124548
    Abstract: The present disclosure relates to techniques for determining optimal routing paths for computing devices in a network, including selecting an optimal gateway among a number of available gateways. The techniques include gathering data regarding characteristics of a network, including gateways and network access nodes (NANs) in at least one access network. The characteristics can include, e.g., supported frequency bands, communication protocols, signal-to-noise ratio, power, signal noise and quality, slicing information, and whether a network vender is a standalone network vendor or a non-standalone network vendor. In one aspect, the characteristics are obtained using the Mobile Broadband Interface Model (MBIM). The characteristics can be used by devices in determining routing paths based on requirements of individual flows and/or workflows of individual application instances.
    Type: Application
    Filed: December 23, 2021
    Publication date: April 21, 2022
    Inventors: Amar Srivastava, Kshitij A. Doshi, Cristian Florin Dumitrescu, Christian Maciocco