Patents by Inventor Hongzhong Jia

Hongzhong Jia has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20230144751
    Abstract: A machine learning model is trained. A feature importance metric is determined for each machine learning feature of a plurality of machine learning features of the machine learning model. Based on the feature importance metrics, one or more machine learning features of the plurality of machine learning features of the machine learning model are managed.
    Type: Application
    Filed: November 14, 2022
    Publication date: May 11, 2023
    Inventors: Hongzhong Jia, Jay Parikh
  • Patent number: 11531831
    Abstract: A machine learning model is trained. A feature importance metric is determined for each machine learning feature of a plurality of machine learning features of the machine learning model. Based on the feature importance metrics, one or more machine learning features of the plurality of machine learning features of the machine learning model are managed.
    Type: Grant
    Filed: September 30, 2019
    Date of Patent: December 20, 2022
    Assignee: Meta Platforms, Inc.
    Inventors: Hongzhong Jia, Jay Parikh
  • Patent number: 11003992
    Abstract: In one embodiment, a method includes establishing access to first and second different computing systems. A machine learning model is assigned for training to the first computing system, and the first computing system creates a check-point during training in response to a first predefined triggering event. The check-point may be a record of an execution state in the training of the machine learning model by the first computing system. In response to a second predefined triggering event, the training of the machine learning model on the first computing system is halted, and in response to a third predefined triggering event, the training of the machine learning model is transferred to the second computing system, which continues training the machine learning model starting from the execution state recorded by the check-point.
    Type: Grant
    Filed: October 16, 2017
    Date of Patent: May 11, 2021
    Assignee: Facebook, Inc.
    Inventors: Lukasz Wesolowski, Mohamed Fawzi Mokhtar Abd El Aziz, Aditya Rajkumar Kalro, Hongzhong Jia, Jay Parikh
  • Publication number: 20210097329
    Abstract: A machine learning model is trained. A feature importance metric is determined for each machine learning feature of a plurality of machine learning features of the machine learning model. Based on the feature importance metrics, one or more machine learning features of the plurality of machine learning features of the machine learning model are managed.
    Type: Application
    Filed: September 30, 2019
    Publication date: April 1, 2021
    Inventors: Hongzhong Jia, Jay Parikh
  • Publication number: 20210097429
    Abstract: It is determined that a selected server among a pool of servers is eligible to be utilized for machine learning training. At least the selected server is utilized to train at least a portion of a machine learning model. It is determined that the selected server among the pool of servers is no longer eligible to be utilized for machine learning training. A training state of the machine learning model is saved. The selected server is returned for other use in the pool of servers.
    Type: Application
    Filed: September 30, 2019
    Publication date: April 1, 2021
    Inventors: Hongzhong Jia, Jay Parikh
  • Patent number: 10545934
    Abstract: A dataset management system (“system”) reduces the amount of data to be stored for future analyses. The system determines a sampling rate of the data based on a required level of accuracy, and samples the data at the determined sampling rate. Initially, all data transactions (“full dataset”) and the sampled data (“sampled dataset”) are logged and stored. Based upon a trigger condition, e.g., after a specified period, the full dataset and the sampled dataset are analyzed separately and the analysis results are compared. If the comparison is sufficiently similar (i.e., the sampling produces a sufficiently accurate set of data or a variance between the analysis results of the datasets is within a specified threshold), the system discontinues full data logging and stores only the sampled dataset. Further, the full dataset is deleted. The sampling thus reduces the required data volume significantly, thereby minimizing consumption of the storage space.
    Type: Grant
    Filed: June 30, 2017
    Date of Patent: January 28, 2020
    Assignee: Facebook, Inc.
    Inventors: Hongzhong Jia, Rajiv Krishnamurthy, Lin Qiao, Joshua David Metzler
  • Patent number: 10379558
    Abstract: Embodiments are described for dynamically responding to demand for server computing resources. The embodiments can monitor performance of each of multiple computing systems in a data center, identify a particular computing system of the multiple computing systems for allocation of additional computing power, determine availability of an additional power supply to allocate to the identified computing system, determine availability of a capacity on a power distribution line connected to the particular computing system to provide the additional power supply to the particular computing system, and allocate the additional computing power to the identified computing system as a function of the determined availability of the additional power supply and the determined availability of the capacity on the power distribution line.
    Type: Grant
    Filed: August 13, 2014
    Date of Patent: August 13, 2019
    Assignee: Facebook, Inc.
    Inventors: Xiaojun Liang, Yusuf Abdulghani, Ming Ni, Hongzhong Jia, Jason Taylor
  • Publication number: 20190182980
    Abstract: The disclosure is directed to placement of server racks of different types in a data center for efficient allocation of resources to the servers. A data center has limited physical resources (e.g., electrical power, cooling, airflow, network bandwidth, weight capacity, etc.). Various server rack types (e.g., hosting a type of a server computer) consume different amounts of these resources. If the distribution of server rack types in a data center is imbalanced, various unexpected failures can occur. The systems considers resource utilizations of all server rack types and generates a deployment layout that assigns these server rack types across multiple rows of the data center to ensure a deployment constraint of the data center is satisfied. Application services that are run on these racks are bucketed based on their resource consumption. Each bucket is distributed in a similar manner as the rack type across the data center.
    Type: Application
    Filed: December 7, 2017
    Publication date: June 13, 2019
    Inventors: Hongzhong Jia, Yusuf Abdulghani, Parth M. Malani
  • Patent number: 10313452
    Abstract: A chat messaging service provided for a chat user is migrated. At a second chat server from a first chat server, static information associated with a chat user is received. At the second chat server from the first chat server, dynamic information associated with the chat user is received. At least a portion of the dynamic information is received after the chat user is indicated as being associated with the migration state. After the chat user is no longer indicated as being associated with the migration state, a chat message for the chat user is received at the second chat server.
    Type: Grant
    Filed: February 14, 2017
    Date of Patent: June 4, 2019
    Assignee: Facebook, Inc.
    Inventors: Hongzhong Jia, Xiaojun Liang, Li Hua, Goranka Bjedov
  • Publication number: 20190114537
    Abstract: In one embodiment, a method includes establishing access to first and second different computing systems. A machine learning model is assigned for training to the first computing system, and the first computing system creates a check-point during training in response to a first predefined triggering event. The check-point may be a record of an execution state in the training of the machine learning model by the first computing system. In response to a second predefined triggering event, the training of the machine learning model on the first computing system is halted, and in response to a third predefined triggering event, the training of the machine learning model is transferred to the second computing system, which continues training the machine learning model starting from the execution state recorded by the check-point.
    Type: Application
    Filed: October 16, 2017
    Publication date: April 18, 2019
    Inventors: Lukasz Wesolowski, Mohamed Fawzi Mokhtar Abd El Aziz, Aditya Rajkumar Kalro, Hongzhong Jia, Jay Parikh
  • Patent number: 10244052
    Abstract: The disclosure is directed to receiving a digitized content item that is indicated to be shared with users (e.g., all, some, or one) of a social networking system; selecting, based on one or more characteristics of the digitized content item, a second server computing device at which to further compute or store the digitized content item; and transmitting, to the second computing device, the digitized content item for storage at the second computing device. At least one of the characteristics can be a classification of a user who transmitted the digitized content item via the data communications network. By selecting second server computing devices based on characteristics, data communications network performance can be improved.
    Type: Grant
    Filed: November 23, 2016
    Date of Patent: March 26, 2019
    Assignee: Facebook, Inc.
    Inventors: Jay Parikh, Hongzhong Jia
  • Publication number: 20190005075
    Abstract: A dataset management system (“system”) reduces the amount of data to be stored for future analyses. The system determines a sampling rate of the data based on a required level of accuracy, and samples the data at the determined sampling rate. Initially, all data transactions (“full dataset”) and the sampled data (“sampled dataset”) are logged and stored. Based upon a trigger condition, e.g., after a specified period, the full dataset and the sampled dataset are analyzed separately and the analysis results are compared. If the comparison is sufficiently similar (i.e., the sampling produces a sufficiently accurate set of data or a variance between the analysis results of the datasets is within a specified threshold), the system discontinues full data logging and stores only the sampled dataset. Further, the full dataset is deleted. The sampling thus reduces the required data volume significantly, thereby minimizing consumption of the storage space.
    Type: Application
    Filed: June 30, 2017
    Publication date: January 3, 2019
    Inventors: Hongzhong Jia, Rajiv Krishnamurthy, Lin Qiao, Joshua David Metzler
  • Publication number: 20180146032
    Abstract: The disclosure is directed to receiving a digitized content item that is indicated to be shared with users (e.g., all, some, or one) of a social networking system; selecting, based on one or more characteristics of the digitized content item, a second server computing device at which to further compute or store the digitized content item; and transmitting, to the second computing device, the digitized content item for storage at the second computing device. At least one of the characteristics can be a classification of a user who transmitted the digitized content item via the data communications network. By selecting second server computing devices based on characteristics, data communications network performance can be improved.
    Type: Application
    Filed: November 23, 2016
    Publication date: May 24, 2018
    Inventors: Jay Parikh, Hongzhong Jia
  • Publication number: 20170214751
    Abstract: A chat messaging service provided for a chat user is migrated. At a second chat server from a first chat server, static information associated with a chat user is received. At the second chat server from the first chat server, dynamic information associated with the chat user is received. At least a portion of the dynamic information is received after the chat user is indicated as being associated with the migration state. After the chat user is no longer indicated as being associated with the migration state, a chat message for the chat user is received at the second chat server.
    Type: Application
    Filed: February 14, 2017
    Publication date: July 27, 2017
    Inventors: Hongzhong Jia, Xiaojun Liang, Li Hua, Goranka Bjedov
  • Patent number: 9697247
    Abstract: The disclosure is directed to storing data in different tiers of a database based on the access pattern of the data. Immutable data, e.g., data that does not change or changes less often than a specified threshold, is stored in a first storage tier of the database, and mutable data, e.g., data that changes more often than immutable data, is stored in a second storage tier of the database. The second storage tier of the database is more performant than the first storage tier, e.g., the second storage tier has a higher write endurance and a lower write latency than the first storage tier. All writes to the database are performed at the second storage tier and reads on both storage tiers. The storage tiers are synchronized, e.g., the set of data is copied from the second to the first storage tier based on a trigger, e.g., a specified schedule.
    Type: Grant
    Filed: July 16, 2014
    Date of Patent: July 4, 2017
    Assignee: Facebook, Inc.
    Inventors: Narsing Vijayrao, Hongzhong Jia, Jason Taylor, Mark Douglas Callaghan, Domas Mituzas
  • Patent number: 9608831
    Abstract: Migrating a chat messaging service provided for a chat user is disclosed. At a second chat server from a first chat server, static information associated with a chat user is received. The static information is received before the chat user is indicated as being associated with a migration state. At the second chat server from the first chat server, dynamic information associated with the chat user is received. At least a portion of the dynamic information is received after the chat user is indicated as being associated with the migration state. After the chat user is no longer indicated as being associated with the migration state, a chat message for the chat user is received at the second chat server.
    Type: Grant
    Filed: June 22, 2012
    Date of Patent: March 28, 2017
    Assignee: Facebook, Inc.
    Inventors: Hongzhong Jia, Xiaojun Liang, Li Hua, Goranka Bjedov
  • Patent number: 9396500
    Abstract: Techniques to adaptively manage service requests within a multi-server system. In one embodiment, a service request and a service rule associated with the service request are received. Data about operating parameters of at least one server in a multi-server system are also received as part of a feedback loop. A response to the service request based on the service rule and the operating parameters is determined. Execution of the service request may be modified according to a tiered service rule based on the at least one server reaching a capacity threshold. The modification includes omitting an action in execution of the service request.
    Type: Grant
    Filed: June 20, 2012
    Date of Patent: July 19, 2016
    Assignee: Facebook, Inc.
    Inventors: Andrew Barkett, Hongzhong Jia, Xiaojun Liang, John Morrow, Daniil Neiter
  • Patent number: 9377958
    Abstract: Channel performance can be improved in a storage device, such as a flash memory or a flash-based solid state drive, by dynamically provisioning available data channels for both write and read operations. In one aspect, a set of available data channels on a storage device is partitioned into a set of write channels and a set of read channels according to a read-to-write ratio. Next, when an incoming data stream of mixed read requests and write requests arrives at the storage device, the allocated read channels process the read requests on a first group of memory blocks, which does not include garbage collection and write amplification on the first group of memory blocks. In parallel, the allocated write channels process the write requests on a second group of memory blocks, which does include garbage collection and write amplification on the second group of memory blocks.
    Type: Grant
    Filed: August 12, 2014
    Date of Patent: June 28, 2016
    Assignee: Facebook, Inc.
    Inventors: Narsing Vijayrao, Hongzhong Jia, Jason Taylor
  • Patent number: 9336155
    Abstract: Storing data in a cache is disclosed. It is determined that a data record is not stored in a cache. A random value is generated using a threshold value. It is determined whether to store the data record in the cache based at least in part on the generated random value.
    Type: Grant
    Filed: August 31, 2015
    Date of Patent: May 10, 2016
    Assignee: Facebook, Inc.
    Inventors: Hongzhong Jia, Xiaojun Liang, Jason Taylor
  • Patent number: 9277026
    Abstract: Various embodiments of the present disclosure relate to a cache stickiness index for providing measurable metrics associated with caches of a content delivery networking system. In one embodiment, a method for generating a cache stickiness index, including a cluster stickiness index and a region stickiness index, is disclosed. In embodiments, the cluster stickiness index is generated by comparing cache keys shared among a plurality of front-end clusters. In embodiments, the region stickiness index is generated by comparing cache keys shared among a plurality of data centers. In one embodiment, a system comprising means for generating a stickiness index is disclosed.
    Type: Grant
    Filed: July 3, 2013
    Date of Patent: March 1, 2016
    Assignee: Facebook, Inc.
    Inventors: Xiaojun Liang, Hongzhong Jia, Jason Taylor