Patents by Inventor Todd Little

Todd Little has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20190104196
    Abstract: In accordance with an embodiment, described herein are systems and methods for implementing a distributed ledger a blockchain cloud service. The blockchain cloud service can include nodes of the distributed ledger and a REST proxy service component. The REST proxy service uses a service development kit for the distributed ledger to communicate with the distributed ledger, and can provide REST APIs for use by client applications to query through chaincodes, synchronously or asynchronously invoke transactions through the chaincodes, get transaction statuses, and get BCS proxy versions. The REST proxy service component can authenticate REST calls, and translate the REST calls into remote procedural calls, for use in interfacing with the distributed ledger. The REST proxy service component can further provide REST APIs that support the same functions which are provided by the BCS management console component, and provide a user interface for client applications to consume the BCS instance.
    Type: Application
    Filed: September 25, 2018
    Publication date: April 4, 2019
    Inventors: Jared Li, Pierce Shi, Shi Xiang Zhou, Weiguo Zhu, Sheng Zhu, Shun Li, Jim Jin, Todd Little
  • Publication number: 20190102423
    Abstract: In accordance with an embodiment, described herein is a system and method for implementing a distributed ledger a blockchain cloud service. The blockchain cloud service can include nodes of the distributed ledger and a management console component. The management console component can include a web application running in a script runtime environment, a plurality of backend of APIs for communicating with various nodes of the blockchain cloud service, and a plurality of client APIs configured to be invoked by a client application. The plurality of client APIs uses one or more of the plurality of backend APIs in provisioning the distributed ledger as a blockchain cloud service, and in managing the managing the blockchain cloud service.
    Type: Application
    Filed: September 25, 2018
    Publication date: April 4, 2019
    Inventors: Todd Little, Pierce Shi, Jared Li, Shi Xiang Zhou, Weiguo Zhu, Sheng Zhu, Shun Li, Jim Jin, Qingsheng Zhang
  • Patent number: 10244068
    Abstract: In accordance with an embodiment, described herein is a system and method for providing distributed caching in a transactional processing environment. The caching system can include a plurality of layers that provide a caching feature for a plurality of data types, and can be configured for use with a plurality of caching providers. A common data structure can be provided to store serialized bytes of each data type, and architecture information of a source platform executing a cache-setting application, so that a cache-getting application can use the information to convert the serialized bytes to a local format. A proxy server can be provided to act as a client to a distributed in-memory grid, and advertise services to a caching client, where each advertised service can match a cache in the distributed in-memory data grid, such as Coherence. The caching system can be used to cache results from a service.
    Type: Grant
    Filed: January 15, 2016
    Date of Patent: March 26, 2019
    Assignee: ORACLE INTERNATIONAL CORPORATION
    Inventors: Todd Little, Xugang Shen, Jim Yongshun Jin, Jesse Hou
  • Publication number: 20190089803
    Abstract: In accordance with an embodiment, described herein is a system and method for conditional call path monitoring in a distributed transactional middleware environment. A cache can be provided in local memory, for use by an agent in the reporting and aggregation of call path metrics. When the agent collects such metrics, it does not report them immediately to a system and application monitor (SAM) manager (e.g., Tuxedo System and Application Monitor, TSAM), but instead stores them in the cache, indexed by correlation ID (identifier). When a predefined condition is met at a participating node, that node propagates a corresponding correlation ID to other participating nodes, via the SAM manager. The other participating nodes can then search for the correlation ID in the cache, and report to the SAM manager metrics of call paths which meet the condition.
    Type: Application
    Filed: May 9, 2018
    Publication date: March 21, 2019
    Inventors: JARED LI, SHENG ZHU, CHRIS GUO, BINGBING LUAN, XIAOOU SUN, TODD LITTLE
  • Patent number: 10198271
    Abstract: A system and method for booting application servers in parallel. In an exemplary method, each server and server group entry can be associated with a dependency attribute. The method can create a dependency map based upon the dependency attributes. Servers that do not have a dependency can be booted in parallel. The remaining servers and server groups can be booted based upon the dependency map.
    Type: Grant
    Filed: January 15, 2016
    Date of Patent: February 5, 2019
    Assignee: ORACLE INTERNATIONAL CORPORATION
    Inventors: Xianzheng Lv, Todd Little, Xiangdong Li, Jim Yongshun Jin
  • Patent number: 10200494
    Abstract: In accordance with an embodiment, described herein is a system and method for providing distributed caching in a transactional processing environment. The caching system can include a plurality of layers that provide a caching feature for a plurality of data types, and can be configured for use with a plurality of caching providers. A common data structure can be provided to store serialized bytes of each data type, and architecture information of a source platform executing a cache-setting application, so that a cache-getting application can use the information to convert the serialized bytes to a local format. A proxy server can be provided to act as a client to a distributed in-memory grid, and advertise services to a caching client, where each advertised service can match a cache in the distributed in-memory data grid, such as Coherence. The caching system can be used to cache results from a service.
    Type: Grant
    Filed: January 15, 2016
    Date of Patent: February 5, 2019
    Assignee: ORACLE INTERNATIONAL CORPORATION
    Inventors: Todd Little, Xugang Shen, Jim Yongshun Jin, Jesse Hou
  • Patent number: 9996403
    Abstract: A middleware machine environment can provide message queues for multinode applications. The transactional middleware machine environment includes a message control data structure on a message receiver and a heap data structure in a shared memory that is associated with the message receiver. The message sender operates to write a message directly into the heap data structure, and to maintain metadata associated with the message in the message control data structure. Furthermore, the message control data structure can be a ring structure with a head pointer and a tail pointer. Additionally, the message receiver resides on a server that is connected with a plurality of clients, with each of said clients keeping a private copy of the message control data structure. Also, the message receiver can support concurrent access to the message control data structure associated with the message receiver.
    Type: Grant
    Filed: August 10, 2012
    Date of Patent: June 12, 2018
    Assignee: ORACLE INTERNATIONAL CORPORATION
    Inventors: Richard Frank, Todd Little, Arun Kaimalettu, Leonard Tominna
  • Patent number: 9894175
    Abstract: In accordance with an embodiment, described herein is a system and method for providing distributed caching in a transactional processing environment. The caching system can include a plurality of layers that provide a caching feature for a plurality of data types, and can be configured for use with a plurality of caching providers. A common data structure can be provided to store serialized bytes of each data type, and architecture information of a source platform executing a cache-setting application, so that a cache-getting application can use the information to convert the serialized bytes to a local format. A proxy server can be provided to act as a client to a distributed in-memory grid, and advertise services to a caching client, where each advertised service can match a cache in the distributed in-memory data grid, such as Coherence. The caching system can be used to cache results from a service.
    Type: Grant
    Filed: January 15, 2016
    Date of Patent: February 13, 2018
    Assignee: ORACLE INTERNATIONAL CORPORATION
    Inventors: Todd Little, Xugang Shen, Jim Yongshun Jin, Jesse Hou
  • Patent number: 9715451
    Abstract: In accordance with an embodiment, described herein is a system and method for providing distributed caching in a transactional processing environment. The caching system can include a plurality of layers that provide a caching feature for a plurality of data types, and can be configured for use with a plurality of caching providers. A common data structure can be provided to store serialized bytes of each data type, and architecture information of a source platform executing a cache-setting application, so that a cache-getting application can use the information to convert the serialized bytes to a local format. A proxy server can be provided to act as a client to a distributed in-memory grid, and advertise services to a caching client, where each advertised service can match a cache in the distributed in-memory data grid, such as Coherence. The caching system can be used to cache results from a service.
    Type: Grant
    Filed: January 15, 2016
    Date of Patent: July 25, 2017
    Assignee: ORACLE INTERNATIONAL CORPORATION
    Inventors: Todd Little, Xugang Shen, Jim Yongshun Jin, Jesse Hou
  • Publication number: 20170116027
    Abstract: In accordance with an embodiment, described herein is a system and method for making requests to a specific server in a transactional processing environment with a multiple servers, single queue (MSSQ) configuration. Each server in a MSSQ set can be associated with a secondary request queue, and can include a main thread, a secondary thread, and an internal memory queue. Each server can use an application programming interface (API) to advertise one or more non-unanimous services on a secondary request queue associated with that server. The secondary thread of a particular server is configured to dequeue, from the secondary request queue associated with the particular server, a request message for a non-unanimous service, and stores the dequeued message in the internal memory queue. The main thread of the particular server checks the internal memory queue before checking the main request queue for messages to process.
    Type: Application
    Filed: January 15, 2016
    Publication date: April 27, 2017
    Inventors: TODD LITTLE, QINGSHENG ZHANG, JIM YONGSHUN JIN
  • Publication number: 20170116121
    Abstract: In accordance with an embodiment, described herein is a system and method for providing distributed caching in a transactional processing environment. The caching system can include a plurality of layers that provide a caching feature for a plurality of data types, and can be configured for use with a plurality of caching providers. A common data structure can be provided to store serialized bytes of each data type, and architecture information of a source platform executing a cache-setting application, so that a cache-getting application can use the information to convert the serialized bytes to a local format. A proxy server can be provided to act as a client to a distributed in-memory grid, and advertise services to a caching client, where each advertised service can match a cache in the distributed in-memory data grid, such as Coherence. The caching system can be used to cache results from a service.
    Type: Application
    Filed: January 15, 2016
    Publication date: April 27, 2017
    Inventors: TODD LITTLE, XUGANG SHEN, JIM YONGSHUN JIN, JESSE HOU
  • Publication number: 20170118300
    Abstract: In accordance with an embodiment, described herein is a system and method for providing distributed caching in a transactional processing environment. The caching system can include a plurality of layers that provide a caching feature for a plurality of data types, and can be configured for use with a plurality of caching providers. A common data structure can be provided to store serialized bytes of each data type, and architecture information of a source platform executing a cache-setting application, so that a cache-getting application can use the information to convert the serialized bytes to a local format. A proxy server can be provided to act as a client to a distributed in-memory grid, and advertise services to a caching client, where each advertised service can match a cache in the distributed in-memory data grid, such as Coherence. The caching system can be used to cache results from a service.
    Type: Application
    Filed: January 15, 2016
    Publication date: April 27, 2017
    Inventors: TODD LITTLE, XUGANG SHEN, JIM YONGSHUN JIN, JESSE HOU
  • Publication number: 20170118284
    Abstract: A self-describing data format capable of carrying payload information, such as Tuxedo payload information, as well as filter information. The data format can allow for expression of some or all of Tuxedo Typed Buffers, including STRING, CARRAY, MBSTRING, VIEW, VIEW32, FML, and FML32. The data format is also capable of supporting nesting and error-checking. The proposal of user payload collection can allow customers to siphon off user/payload data to be used in applications, such as Business Intelligence applications, without the need for additional information to be passed with the payload data.
    Type: Application
    Filed: January 15, 2016
    Publication date: April 27, 2017
    Inventors: HUI CHEN, TODD LITTLE, XIANGDONG LI, JIM YONGSHUN JIN
  • Publication number: 20170115995
    Abstract: A system and method for booting application servers in parallel. In an exemplary method, each server and server group entry can be associated with a dependency attribute. The method can create a dependency map based upon the dependency attributes. Servers that do not have a dependency can be booted in parallel. The remaining servers and server groups can be booted based upon the dependency map.
    Type: Application
    Filed: January 15, 2016
    Publication date: April 27, 2017
    Inventors: Xianzheng LV, TODD LITTLE, Xiangdong LI, Jim Yongshun JIN
  • Publication number: 20170116123
    Abstract: In accordance with an embodiment, described herein is a system and method for providing distributed caching in a transactional processing environment. The caching system can include a plurality of layers that provide a caching feature for a plurality of data types, and can be configured for use with a plurality of caching providers. A common data structure can be provided to store serialized bytes of each data type, and architecture information of a source platform executing a cache-setting application, so that a cache-getting application can use the information to convert the serialized bytes to a local format. A proxy server can be provided to act as a client to a distributed in-memory grid, and advertise services to a caching client, where each advertised service can match a cache in the distributed in-memory data grid, such as Coherence. The caching system can be used to cache results from a service.
    Type: Application
    Filed: January 15, 2016
    Publication date: April 27, 2017
    Inventors: TODD LITTLE, XUGANG SHEN, JIM YONGSHUN JIN, JESSE HOU
  • Publication number: 20170116122
    Abstract: In accordance with an embodiment, described herein is a system and method for providing distributed caching in a transactional processing environment. The caching system can include a plurality of layers that provide a caching feature for a plurality of data types, and can be configured for use with a plurality of caching providers. A common data structure can be provided to store serialized bytes of each data type, and architecture information of a source platform executing a cache-setting application, so that a cache-getting application can use the information to convert the serialized bytes to a local format. A proxy server can be provided to act as a client to a distributed in-memory grid, and advertise services to a caching client, where each advertised service can match a cache in the distributed in-memory data grid, such as Coherence. The caching system can be used to cache results from a service.
    Type: Application
    Filed: January 15, 2016
    Publication date: April 27, 2017
    Inventors: Todd Little, Xugang Shen, Jim Yongshun Jin, Jesse Hou
  • Patent number: 9558048
    Abstract: A middleware machine environment can manage message queues for multimode applications. The middleware machine environment includes a shared memory on a message receiver, wherein the shared memory maintains one or more message queues for the middleware machine environment. The middleware machine environment further includes a daemon process that is capable of creating at least one message queue in the shared memory, when a client requests that the at least one message queue be set up to support sending and receiving messages. Additionally, different processes on a client operate to use at least one proxy to communicate with the message server. Furthermore, the middleware machine environment can protect message queues for multimode applications using a security token created by the daemon process.
    Type: Grant
    Filed: August 10, 2012
    Date of Patent: January 31, 2017
    Assignee: ORACLE INTERNATIONAL CORPORATION
    Inventors: Richard Frank, Todd Little, Arun Kaimalettu, Leonard Tominna
  • Patent number: 9116761
    Abstract: A transactional middleware system can exchange messages between a local machine and a remote machine using Remote Direct Memory Access (RDMA) protocol to achieve short latency in a manner like a local message transfer. The transactional middleware machine environment can prevent single-point bottleneck and achieve short latency. The transactional middleware machine environment comprises a first message queue associated with a server in a first transactional machine, wherein the server operates to be accessed using a queue address in the first message queue. The transactional middleware machine environment further comprises a second message queue associated with a client in a second transactional machine. The client operates to send a service request message to the server directly using the first message queue, and the server operates to receive the service request message from the first message queue and send a service response message directly to the client using the second message queue.
    Type: Grant
    Filed: March 8, 2012
    Date of Patent: August 25, 2015
    Assignee: ORACLE INTERNATIONAL CORPORATION
    Inventors: Todd Little, Yongshun Jin, Erli Niu
  • Patent number: 9110851
    Abstract: A transactional system can utilize the distributed storage and high availability (HA) capability provided by a clustered database to support easy and feasible disaster recovery. The transactional middleware machine environment comprises one or more transactional application servers associated with a transaction. The one or more transactional application servers operate to persist transactional log information associated with the transaction in a database that connects with said one or more transactional application servers at a local site. The database at the local site operates to replicate the persisted transactional log information to a remote database at a remote site. The remote database allows a different transactional application server at the remote site to recover the persisted transactional log information and complete the transaction, when a disaster disables the local site.
    Type: Grant
    Filed: March 7, 2012
    Date of Patent: August 18, 2015
    Assignee: ORACLE INTERNATIONAL CORPORATION
    Inventors: Todd Little, Xiangdong Li, Xianzheng Lv
  • Publication number: 20130086148
    Abstract: A transactional middleware system can exchange messages between a local machine and a remote machine using Remote Direct Memory Access (RDMA) protocol to achieve short latency in a manner like a local message transfer. The transactional middleware machine environment can prevent single-point bottleneck. The transactional middleware machine environment comprises a first message queue associate with a server in a first transactional machine, wherein the server operates to be accessed using a queue address in the first message queue. The transactional middleware machine environment further comprises a second message queue associated with a client in a second transactional machine. The client operates to send a service request message to the server directly using the first message queue, and the server operates to receive the service request message from the first message queue and send a service response message directly to the client using the second message queue.
    Type: Application
    Filed: March 8, 2012
    Publication date: April 4, 2013
    Applicant: ORACLE INTERNATIONAL CORPORATION
    Inventors: Todd Little, Yongshun Jin, Erli Niu