Data Processing Method and Mobile Device
Performing data processing on a mobile device, where the mobile device includes a first display, a second display and a second touch panel; the first display is configured to: display a home screen which comprises an icon of a first application; display iconic controls of a first application in response to detecting a first gesture on the icon of the first application; the second display is configured to: display a first GUI of the first application in response to detecting the first gesture; the second touch panel is configured to: detect a second gesture on the second touch panel; display a second GUI of the first application on the second display in response to detecting the second gesture, wherein maintaining display iconic controls of the first application on the first display when displaying the second GUI on the second display.
This is a continuation of U.S. patent application Ser. No. 17/130,317 filed on Dec. 22, 2020, which is a continuation of U.S. patent application Ser. No. 16/481,417 filed on Jul. 26, 2019, now U.S. Pat. No. 10,908,868, which is a national stage entry of Int'l Patent App. No. PCT/CN2017/075633, filed on Mar. 3, 2017, which claims priority to Int'l Patent App. No. PCT/CN2017/073173, filed on Feb. 9, 2017 and Chinese Patent App. No. 201710061746.7, filed on Jan. 26, 2017, all of which are incorporated by reference.
FIELDThe disclosure relates to the field of data processing, and in particular, to a data processing method and a mobile device.
BACKGROUNDWith popularity of the mobile Internet, a growing quantity of users are accustomed to reading ebooks on mobile devices (such as mobile phones, tablet computers, and ebook readers). When reading an ebook on a mobile device, a user needs to perform page flip. Page flip is usually implemented by the user by touching a screen of the mobile device with a finger, to turn to a next page or return to a previous page. To enable the ebook to accurately jump to a particular chapter that the user desires to read, the user needs to perform many complex operations. As shown in
To resolve the foregoing technical problem, embodiments provide a data processing method and a mobile device, to reduce operation steps of a user, improve transaction processing efficiency of a mobile device, and further enhance user experience.
According to a first aspect, an embodiment provides a method for performing data processing on a mobile device. The mobile device includes a touch display screen disposed on a front face of the mobile device, and a low power display screen and a touch panel that are disposed on a rear face of the mobile device, the touch display screen is a capacitive touchscreen, and the low power display screen is an electronic ink screen. The method includes: detecting, by the touch display screen, a touch operation of a user; determining, by the mobile device in response to the touch operation, whether the touch operation works on an application program of a reading type; when the mobile device determines that the touch operation works on the application program of a reading type, instructing, by the mobile device, the low power display screen to display a first graphical user interface (first GUI) of the application program, and instructing to turn off a power supply of the touch display screen and activate the touch panel, where the first GUI is an interface, in the application program, for displaying content of an ebook; and when the touch panel detects a touch event of the user, determining, by the mobile device based on a speed and a direction of the gesture, an instruction corresponding to the gesture, and executing the instruction, where a result of executing the instruction is: displaying a second graphical user interface (second GUI) of the application program on the low power display screen based on the speed and the direction of the gesture, where the second GUI is used to display a catalogue of the ebook. In this way, the second touch display screen and the touch panel are added on the rear face of the mobile device, and the touch event of the user on the touch panel is detected to execute the corresponding instruction, so that different graphical user interfaces are displayed for the ebook displayed on the second touch display screen. This simplifies operation steps when the user reads the ebook, and also improves transaction processing efficiency of the mobile device.
In some embodiments, the displaying a second graphical user interface of the application program on the low power display screen may be specifically: displaying a part of the catalogue of the ebook on the low power display screen based on the speed and the direction of the gesture.
According to a second aspect, an embodiment provides a method for performing data processing on a mobile device. The mobile device includes a first touch display screen disposed on a front face of the mobile device, and a second touch display screen and a touch panel that are disposed on a rear face of the mobile device. The method includes: when the first touch display screen detects a touch operation of a user, determining, by the mobile device, whether the touch operation works on an application program; when determining that the touch operation works on the application program, determining, by the mobile device, a type of the application program; when the mobile device determines that the type of the application program is reading, instructing, by the mobile device, the second touch display screen to display a first graphical user interface of the application program; and when the touch panel detects a touch event of the user, determining, by the mobile device based on the touch event, an instruction corresponding to the touch event, and executing the instruction, where a result of executing the instruction is: displaying a second graphical user interface of the application program on the second touch display screen based on the touch event.
In some embodiments, the determining, by the mobile device based on the touch event, an instruction corresponding to the touch event, and executing the instruction may be specifically: detecting, by the touch panel at different time points t1, t2, and t3, that positions of three touch points A, B, and C in the touch event are (x1, y1), (x2, y2), and (x3, y3); determining, by the mobile device, a speed and/or a direction of the touch event based on X-axis coordinate values x1, x2, and x3 of the three touch points A, B, and C and the time points t1, t2, and t3; and performing, by the mobile device based on the speed and/or the direction, a fast page flip operation on content of an ebook displayed on the first graphical user interface, and displaying the second graphical user interface of the fast page flip operation on the second touch display screen.
In some embodiments, the second graphical user interface may be used to display a catalogue of the ebook.
In some other embodiments, the touch event is a gesture of sliding on the touch panel by a single finger, and the displaying a second graphical user interface of the application program on the second touch display screen based on the touch event may be specifically: displaying a fast page flip graphical user interface of the application program on the second touch display screen based on a speed and a direction of the sliding gesture.
In some other embodiments, a processor 380 may alternatively instruct a first display panel 341 to display some graphical user interfaces of the application program, and other graphical user interfaces are displayed on a second display panel 342. In this way, a user can focus more on content of interest, and a screen splitting function can be implemented. It may be understood that the some graphical user interfaces displayed on the first display panel 341 may be specifically iconic controls of the application program and the other graphical user interfaces displayed on the second display panel 342 may mainly be substantial content of the application program. For example, if the application program is “reading”, content of an ebook that is being read is displayed on the second display panel 342, and some controls such as “page up” and “page down” that are not related to substantial content of the ebook are displayed on the first display panel 341.
In some other embodiments, the mobile device further turns off a power supply of the first touch display screen while the mobile device instructs the second touch display screen to display the first graphical user interface of the application program.
According to a third aspect, an embodiment provides a mobile device for performing data processing. The mobile device includes a first display panel and a first touch panel that are disposed on a front face of the mobile device, and a second display panel and a second touch panel that are disposed on a rear face of the mobile device; and the mobile device further includes a processor, a memory, and a power supply management system. The first display panel is configured to display an icon, a graphical user interface, and a component of an application program; when the first touch panel detects a touch operation of a user, the processor determines whether the touch operation works on application program displayed on the first display panel; when the processor determines that the touch operation works on the application program, the processor determines a type of the application program; when the processor determines that the type of the application program is reading, the processor instructs the second display panel to display a first graphical user interface of the application program, and instructs the power supply management system to turn off a power supply of the first display panel and activate the second display panel, where the first graphical user interface is an interface, in the application program, for displaying content of an ebook; and when the second touch panel detects a touch event of the user, the processor determines, based on a speed and a direction of the touch event, an instruction corresponding to the touch event, and executes the instruction, where a result of executing the instruction is: displaying a second graphical user interface of the application program on the second display panel based on the speed and the direction of the touch event, where the second graphical user interface is used to display a catalogue of the ebook.
In some embodiments, the first display panel is a liquid crystal display, and the second display panel is an electronic ink screen. In this way, the first display panel may be mainly configured for daily operations of the user, and the second display panel is a display panel of a low power material, and therefore may be configured to read an ebook.
In some other embodiments, the mobile device may be a mobile phone or a tablet computer.
In the foregoing embodiments, the touch operation of the user on the touch panel is detected, to determine to perform processing (for example, page flip or returning to the catalogue) on the content of the ebook displayed on the second display panel. This greatly simplifies operation steps of the user, improves transaction processing efficiency of the mobile device, and further enhances user experience.
According to a fourth aspect, an embodiment provides a method for performing data processing on a mobile device. The mobile device includes a first touch display screen disposed on a front face of the mobile device, and a second touch display screen and a touch panel that are disposed on a rear face of the mobile device, the first touch display screen may be a capacitive touchscreen, and the second touch display screen may be an electronic ink screen. The method may specifically include: detecting a touch operation of a user on the first touch display screen; displaying, in response to the touch operation, a control for an application program on the first touch display screen; and in response to a touch operation performed by the user on the control, displaying a first graphical user interface of the application program on the first touch display screen, and displaying a second graphical user interface of the application program on the second touch display screen, where the first graphical user interface includes a plurality of controls, the second graphical user interface displays content of the application program, and the plurality of controls are used to control the content of the application program.
In some other embodiments, the application program is an application program of a reading type, and the content of the application program may be an ebook.
According to a fifth aspect, an embodiment provides a method for performing data processing on a mobile device. The mobile device includes a first touch display screen disposed on a front face of the mobile device, and a second touch display screen and a touch panel that are disposed on a rear face of the mobile device, the first touch display screen is a capacitive touchscreen, and the second touch display screen is an electronic ink screen. The method includes: detecting a touch operation of a user on the first touch display screen; displaying, in response to the touch operation, a control for an application program on the first touch display screen; and in response to a touch operation performed by the user on the control, displaying a first graphical user interface of the application program on the first touch display screen, and displaying a second graphical user interface of the application program on the second touch display screen, where the first graphical user interface includes a plurality of controls, the second graphical user interface displays content of the application program, and the plurality of controls are used to control the content of the application program.
In some other embodiments, the application program is an application program of a reading type, and the content of the application program is an ebook.
It should be understood that descriptions about the technical features, technical solutions, and advantages or similar descriptions in this specification do not indicate that all the features and advantages may be implemented in any single embodiment. In contrast, it may be understood that descriptions about the features or advantages mean that a particular technical feature, technical solution, or advantage is included in at least one embodiment. Therefore, the descriptions about the technical features, technical solutions, or advantages in this specification do not necessarily indicate a same embodiment. Further, the technical features, technical solutions and advantages described in the following embodiments may be combined in any proper manner. A person skilled in the art should understand that an embodiment may be implemented without one or more particular technical features, technical solutions or advantages.
A mobile device in the following embodiments may be any device having a wireless communication function, for example, may be a wearable electronic device (for example, a smartwatch) having the wireless communication function, may be a mobile phone 300 shown in
In the following embodiment, a mobile phone is used as an example to describe how a mobile device implements a specific technical solution in this embodiment. As shown in
It should be understood that the mobile phone 300 in the figure is merely an example of the mobile device, and the mobile phone 300 may have more or fewer components than those shown in the figure, a combination of two or more components, or components disposed differently. The components shown in the figure may be implemented by hardware, software, or a combination of hardware and software, where the hardware, the software, or the combination of hardware and software includes one or more signal processing and/or application-specific integrated circuits.
As shown in
The following specifically describes each component of the mobile phone 300 with reference to
The RF circuit 310 may be configured to receive and transmit information, or receive and send signals in a call process. The RF circuit 310 may receive downlink information of a base station and then provide the received downlink information for the processor 380 for processing, and send uplink data to the base station. Usually, the RF circuit includes but is not limited to devices such as an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, and a duplexer. In addition, the RF circuit 310 may further communicate with a network and another mobile device through wireless communication. The wireless communication may use any communications standard or protocol, including but not limited to a Global System for Mobile Communications, a general packet radio service, code division multiple access, wideband code division multiple access, Long Term Evolution, an email, a short message service, and the like.
The memory 320 may be configured to store a software program and data. The processor 380 runs the software program and the data that are stored in the memory 320, to perform various functions of the mobile phone 300 and process data. The memory 320 may mainly include a program storage area and a data storage area. The program storage area may store an operating system, an application program required by at least one function (for example, a sound playback function and an image playback function), and the like. The data storage area may store data (for example, audio data and a phone book) created based on use of the mobile phone 300, and the like. In addition, the memory 320 may include a high-speed random access memory, and may further include a non-volatile memory, for example, at least one disk storage device, a flash memory device, or another volatile solid-state storage device. In the following embodiments, the memory 320 stores an operating system that enables the mobile phone 300 to run, for example, an iOS® operating system developed by Apple Inc., an Android® open source operating system developed by Google Inc., and a Windows® operating system developed by Microsoft Corporation.
The input unit 330 (namely, a touchscreen) may be configured to receive entered numeral or character information, and generate signal input related to user setting and function control of the mobile phone 300. Specifically, the input unit 330 may include a first touch panel 331 disposed on a front face of the mobile phone 300. The first touch panel 331 is also referred to as a first touchscreen, and may collect a touch operation of a user on or near the first touch panel 331 (such as an operation performed by the user on the first touch panel 331 or near the first touch panel 331 by using a finger or any proper object or accessory such as a stylus), and drive a corresponding connection apparatus according to a preset program. Optionally, the first touch panel 331 may include two parts: a touch detection apparatus and a touch controller (not shown in the figure). The touch detection apparatus detects a touch orientation of the user, detects a signal brought by the touch operation, and sends the signal to the touch controller. The touch controller receives touch information from the touch detection apparatus, converts the touch information into touch point coordinates, and sends the touch point coordinates to the processor 380, and can receive and execute an instruction sent by the processor 380. In addition, the first touch panel 331 may be implemented in various types, such as a resistive type, a capacitive type, an infrared type, and a face acoustic wave type. In addition to the first touch panel 331, the input unit 330 may further include a second touch panel 332 (also referred to as a second touchscreen) disposed on a rear face of the mobile phone 300. In other words, the first touch panel 331 and the second touch panel 332 are disposed on two opposite faces of the mobile phone 300. Specifically, the second touch panel 332 may use a same structure as the first touch panel 331. Details are not described herein again.
The display unit 340 (namely, a display screen) may be configured to display information entered by the user or information provided for the user, and graphical user interfaces (referred to as GUIs in the following) of various menus of the mobile phone 300. The display unit 340 may include a first display panel 341 (also referred to as a first display screen) disposed on the front face of the mobile phone 300, and a second display panel 342 (also referred to as a second display screen) disposed on the rear face of the mobile phone 300. In other words, the first display panel 341 and the second display panel 342 are disposed on the two opposite faces of the mobile phone 300. The first display panel 341 may be configured by using a liquid crystal display, an organic light-emitting diode, or a like form. The second display panel 342 may be a screen made by using an electronic paper display technology or another low power display material, for example, an electronic ink screen. Therefore, the second display panel 342 may be used to read an ebook, a magazine, and the like. Certainly, the second display panel 342 may alternatively use a same display material as the first display panel 341.
As shown in
In some other embodiments, pressure sensors may be further disposed on the first touch panel 331, the second touch panel 332, and the third touch panel 333. In this way, when the user performs a touch operation on the foregoing touch panels, the touch panels can further detect a pressure of the touch operation, so that the mobile phone 300 can more accurately detect the touch operation.
The mobile phone 300 may further include at least one sensor 350, for example, a light sensor, a motion sensor, and another sensor. Specifically, the light sensor may include an ambient light sensor and a proximity sensor. As shown in
The audio frequency circuit 360, a speaker 361, and a microphone 362 may provide an audio interface between the user and the mobile phone 300. The audio frequency circuit 360 may transmit, to the speaker 361, a signal converted from received audio data, and the speaker 361 converts the signal into a sound signal for output. In addition, the microphone 362 converts a collected sound signal into an electrical signal, and the audio frequency circuit 360 receives the electrical signal, converts the electrical signal into audio data, and then outputs the audio data to the RF circuit 310, to send the audio data to, for example, another mobile phone, or outputs the audio data to the memory 320 for further processing.
Wi-Fi is a short-range wireless transmission technology. By using the Wi-Fi module 370, the mobile phone 300 may allow the user to receive and send an email, browse a web page, access streaming media, and the like. The Wi-Fi module 370 provides wireless broadband Internet access for the user.
The processor 380 is a control center of the mobile phone 300, and is connected to all components of the entire mobile phone by using various interfaces and lines. The processor 380 runs or executes the software program stored in the memory 320 and invokes the data stored in the memory 320, to perform the functions of the mobile phone 300 and process data, so as to perform overall monitoring on the mobile phone. In some embodiments, the processor 380 may include one or more processing units. An application processor and a modem processor may be further integrated in the processor 380. The application processor primarily processes an operating system, a user interface, an application program, and the like, and the modem processor primarily processes wireless communication. It may be understood that the modem processor may not be integrated into the processor 380.
A Bluetooth module 381 is configured to exchange information with another device by using Bluetooth, a short-range communications protocol. For example, the mobile phone 300 may establish, by using the Bluetooth module 381, a Bluetooth connection to a wearable electronic device (for example, a smartwatch) that also has a Bluetooth module, to exchange data.
The mobile phone 300 further includes the power supply 390 (for example, a battery) that supplies power to each component. The power supply may be logically connected to the processor 380 by using a power supply management system, to implement functions such as charging management, discharging management, and power consumption management by using the power supply management system. It may be understood that, in the following embodiments, the power supply 390 may include two power supplies: One power supply is mainly configured to supply power to the first display panel 341 and the first touch panel 331, and the other power supply is mainly configured to supply power to the second display panel 342, the second touch panel 332, and the third touch panel 333.
The mobile phone 300 may further include a front-facing camera 353, a rear-facing camera 354, and the like. Details are not described herein.
All methods in the following embodiments may be implemented in the mobile phone 300 having the foregoing hardware structure.
As shown in
In some embodiments, when a user touches or approaches the icon 401 on the first display panel 341 by using a finger 402, the first touch panel 331 of the mobile device 400 detects a touch event on or near the first touch panel 331, and transfers the touch event to the processor 380, to determine an instruction corresponding to the touch event. After that, the processor 380 invokes, according to the instruction, an application program stored in a memory 320, and instructs the first display panel 341 to display a graphical user interface (GUI) of the application program (reading), so that the user can perform a specific operation.
It may be understood that the processor 380 may also instruct the second display panel 342 to display the graphical user interface of the application program. In other words, when the user touches the icon 401 on the front face of the mobile device 400, the GUI of the application program (reading) corresponding to the icon 401 is displayed on the second display panel 342 on the rear face of the mobile device 400, so that the user performs reading on the low power second display panel 342. This also saves a power supply of the mobile device 400. In some other embodiments, the processor 380 may alternatively instruct the first display panel 341 to display some GUIs of the application program, and other GUIs are displayed on the second display panel 342. In this way, the user can focus more on content of interest, and a screen splitting function can be implemented. It may be understood that the some GUIs displayed on the first display panel 341 may be specifically iconic controls of the application program and the other GUIs displayed on the second display panel 342 may mainly be substantial content of the application program. For example, as shown in
In some other embodiments, when the GUI of the application program is displayed on the second display panel 342 on the rear face of the mobile device 400, the processor 380 may send, to a power supply management system in a power supply 390, an instruction for interrupting a power supply of the first display panel 341. After receiving the instruction, the power supply management system turns off the power supply of the first display panel 341, so that the first display panel 341 is in a screen-off state. Alternatively, the processor 380 may send, to the power supply management system, an instruction for making the first display panel 341 sleep. After receiving the instruction, the power supply management system makes the first display panel 341 sleep, so that the first display panel 341 is in a screen-off state. This can further save the power supply of the mobile device 400.
In some other embodiments, the mobile device may alternatively display a GUI of the application program on the second display panel 342 only when determining that the application program is an application program for reading an ebook or the like. In other words, the second display panel 342 may allow displaying of only an application program of a specific type, for example, an application program of a reading type. The mobile device may determine a type of an application program in a plurality of manners, for example, determines a type of an application program based on a category attribute (for example, a game, reading, or communication) of the application program. In this embodiment of this application, the mobile device displays, only when determining that an application program corresponding to an icon tapped by a user is an application program of the reading type, a GUI of the application program on the second display panel 342. If determining that the application program corresponding to the icon tapped by the user is not an application program of the reading type, the mobile device displays the GUI of the application program on the first display panel 341, to facilitate a further operation of the user.
When there is content (for example, the GUI of the “reading” application program) displayed on the second display panel 342, the third touch panel 333 is activated, to receive a touch operation of the user. That the third touch panel 333 is activated may be specifically: the third touch panel 333 is powered on by the power supply 390 and can detect the touch operation of the user on or near the third touch panel 333. It may be understood that the third touch panel 333 may alternatively be activated only when a particular GUI is displayed on the second display panel 342. For example, the third touch panel 333 is activated only when the second display panel 342 displays content of a particular chapter of an ebook (for example, chapters displayed in
As shown in
Generally, the touch event may be a rightward (or leftward) sliding gesture or the like made by the user on the third touch panel 333. After collecting the touch event, the third touch panel 333 sends data related to the touch event to the processor 380. Based on the received data, the processor 380 determines and invokes the instruction corresponding to the touch event. If the instruction corresponding to the touch event is an instruction used for fast page flip, a result of executing the instruction is: fast turning the ebook displayed on the second display panel 342 to a particular page and displaying the particular page. In some embodiments, the processor 380 may further calculate a speed of the touch event based on the related data of the touch event, and then determine, based on the speed, a specific speed for performing page flip on the displayed ebook. For example, a faster speed of a sliding gesture leads to a faster page flip speed, in other words, more pages are flipped. On the contrary, a slower speed of a sliding gesture leads to a slower page flip speed, in other words, fewer pages are flipped. In some other embodiments, the processor 380 may alternatively calculate a sliding track length of the touch event on the third touch panel 333 based on the related data of the touch event, and determine, based on the calculated length, a specific speed for performing page flip on the displayed ebook. For example, a longer sliding track length of a sliding gesture leads to a faster page flip speed, in other words, more pages are flipped.
As shown
In some other embodiments, the processor 380 may first determine, based on the touch event, a gesture corresponding to the touch event, and then execute a corresponding instruction based on the determined gesture.
Specifically, the touch event may be a flick gesture, that is, slightly tapping a touch panel (for example, the first touch panel 331 or the second touch panel 332) with a single finger, then quickly sliding, and then quickly leaving the touch panel, for example, scrolling a screen up or down or switching to a left or right picture. The touch event may alternatively be a slide gesture, that is, slightly tapping a touch panel with a single finger, keeping the finger in contact with the touch panel, and then moving, for example, sliding to unlock. The touch event may alternatively be a swipe gesture, that is, touching a touch panel with a plurality of fingers, keeping the fingers in contact with the touch panel, and then moving, for example, pinching with three fingers to return to a home screen. The touch event may alternatively be a tap gesture, that is, slightly tapping a touch panel with a single finger and then leaving the touch panel immediately. The touch event may alternatively be a double tap gesture, that is, performing a tap gesture operation twice within an extremely short time. The touch event may alternatively be a touch & hold gesture, that is, slightly tapping a touch panel with a finger and keeping the finger resting on the touch panel. The touch event may alternatively be a drag gesture, that is, slightly tapping a touch panel with a finger and slowly moving the finger without leaving the touch panel (usually, to a determined target position, for example, dragging a file to a trash box to delete the file). The touch event may alternatively be a pinch gesture, that is, pinching on a touch panel with two fingers (usually a thumb and an index finger). The touch event may alternatively be an unpinch gesture, that is, stretching on a touch panel with two fingers (usually a thumb and an index finger). It may be understood that, in addition to the foregoing listed gestures, the touch event may be further a gesture in another form. A form of the touch event is not limited in this embodiment.
In some other embodiments, a pressure sensor may be further disposed on the third touch panel 333. In this way, the third touch panel 333 can detect a pressure imposed by the user on the touch panel, so that the processor 380 performs more complex and accurate processing. For example, the mobile device may determine a page flip direction based on a gesture direction, and then determine, based on a pressure value of a touch event, a specific quantity of pages that need to be flipped; and may also determine a change of a page flip speed based on a pressure value change, and display different animation effects on the second display panel 342.
The mobile device may determine the gesture of the touch event by using different technical solutions. Specifically, the third touch panel 333 may collect a touch position of a finger on the third touch panel 333 once every 10 milliseconds (ms) after a touch by the user finger is detected, and send collected related data to the processor 380. The processor 380 determines a gesture of the touch event based on the collected related data. For example, the third touch panel 333 determines the gesture of the touch event based on collected different touch positions of the finger at different time points. It may be understood that, if the pressure sensor is disposed on the third touch panel 333, the third touch panel 333 not only can detect touch positions of the user on the third touch panel 333 at different time points, but also can detect pressures imposed by the user on the third touch panel 333 at different time points. In this way, the mobile device 400 can more accurately determine the gesture of the touch event by using parameters collected by the third touch panel 333, such as a time point, a touch position, and a pressure. It may be understood that the processor 380 of the mobile device 400 may further determine a gesture speed based on the foregoing related data detected by the third touch panel 333.
The processor 380 may invoke and execute different instructions based on different gestures. For example, when determining that the touch event is the flick gesture, the processor 380 invokes an instruction corresponding to the flick gesture. A result of executing the instruction is fast flipping pages of the ebook displayed on the second display panel 342 (referred to as a “fast page flip” operation), for example, fast flipping the ebook to page 8 from page 1. It may be understood that a specific page flip speed of the page flip operation may depend on a sliding speed of the flick gesture on the third touch panel 333. As shown in
In some embodiments, when the processor 380 determines that the touch event is the swipe gesture, the processor 380 invokes an instruction corresponding to the swipe gesture. A result of executing the instruction is enabling the ebook displayed on the second display panel 342 to return to a catalogue page from a specific content page. As shown in
In some embodiments, when the processor 380 determines that the touch event is the pinch gesture, the processor 380 invokes an instruction corresponding to the pinch gesture. A result of executing the instruction may be zooming in, to a given multiple, text of specific content of the ebook displayed on the second display panel 342, to enhance a display effect of the ebook. On the contrary, when the processor 380 determines that the touch event is the unpinch gesture, the processor 380 invokes an instruction corresponding to the unpinch gesture. A result of executing the instruction may be zooming out, to a given level, text of specific content of the ebook displayed on the second display panel 342, to reduce a display effect of the ebook and further save power of the mobile device.
In some other embodiments, the second touch panel 332 is further disposed on the rear face of the mobile device. The user may directly perform, on the second touch panel 332, a simple touch operation on an icon displayed on the second display panel 342.
In some other embodiments, the mobile device (for example, a mobile phone) may be used with a wearable device (for example, a smartwatch) paired with the mobile device, to complete the technical solution in the foregoing embodiments. For example, the third touch panel 333 may not be disposed on the mobile phone 300, but a third touch panel is disposed on the smartwatch. In this way, the user may perform a touch gesture operation on the third touch panel of the smartwatch. The touch gesture operation is compiled into an instruction, and the instruction is sent to the mobile device by using a short-range communications protocol (for example, Wi-Fi or Bluetooth) or the like. After receiving the instruction, the mobile device executes the instruction, so that the second display panel 342 on the rear face of the mobile device displays the specific content of the ebook. It may be understood that the mobile device may execute different instructions based on different gestures in the smartwatch, as described in the foregoing embodiments. Details are not described herein.
As shown in
The first display panel 901 is configured to display an icon, a graphical user interface, and a widget of an application program stored in the memory 906.
When the first touch panel 902 detects a touch operation of a user, the processor 905 determines whether the touch operation works on an application program displayed on the first display panel 901.
When the processor 905 determines that the touch operation works on the application program, the processor 905 determines a type of the application program.
When the processor 905 determines that the type of the application program is reading, the processor 905 instructs the second display panel 903 to display a first graphical user interface (first GUI) of the application program, and instructs the power supply management system 907 to turn off a power supply of the first display panel and activate the second display panel 903, where the first GUI is an interface, in the application program, for displaying content of an ebook.
When the second touch panel 904 detects a touch event of the user, the processor 905 determines, based on a speed and a direction of the touch event, an instruction corresponding to the touch event, and executes the instruction.
A result of executing the instruction is: displaying a second graphical user interface (second GUI) of the application program on the second display panel 903 based on the speed and the direction of the touch event, where the second GUI is used to display a catalogue of the ebook.
The first display panel 901 may be a liquid crystal display, and the second display panel 903 may be an electronic ink screen.
In some other embodiments, after it is detected that a user taps an application program “reading” (as shown in
According to the context, the term “if” used in the embodiments may be interpreted as a meaning of “when” or “after” or “in response to determining” or “in response to detecting”. Similarly, according to the context, the phrase “if it is determined that” or “if (a stated condition or event) is detected” may be interpreted as a meaning of “when it is determined that” or “in response to determining” or “when (a stated condition or event) is detected” or “in response to detecting (a stated condition or event)”.
The terms used in the embodiments are merely for the purpose of illustrating specific embodiments, and are not intended to limit this application. The terms “a”, “said” and “the” of singular forms used in the embodiments and the appended claims are also intended to include plural forms, unless otherwise specified in the context clearly. It should also be understood that the term “and/or” used herein indicates and includes any or all possible combinations of one or more associated listed items.
All or some of the foregoing embodiments may be implemented by using software, hardware, firmware, or any combination thereof. When software is used to implement the embodiments, the embodiments may be implemented completely or partially in a form of a computer program product. The computer program product includes one or more computer instructions. When the computer program instructions are loaded and executed on the computer, the procedure or functions according to the embodiments are all or partially generated. The computer may be a general-purpose computer, a dedicated computer, a computer network, or another programmable apparatus. The computer instructions may be stored in a computer-readable storage medium or may be transmitted from a computer-readable storage medium to another computer-readable storage medium. For example, the computer instructions may be transmitted from a website, computer, server, or data center to another website, computer, server, or data center in a wired (for example, a coaxial cable, an optical fiber, or a digital sub scriber line) or wireless (for example, infrared, radio, and microwave, or the like) manner. The computer-readable storage medium may be any usable medium accessible by a computer, or a data storage device, such as a server or a data center, integrating one or more usable media. The usable medium may be a magnetic medium (for example, a floppy disk, a hard disk, or a magnetic tape), an optical medium (for example, a DVD), a semiconductor medium (for example, a solid state disk), or the like.
For a purpose of explanation, the foregoing descriptions are provided with reference to a specific embodiment. However, the foregoing example discussion is not intended to be detailed, and is not intended to limit this application to a disclosed precise form. According to the foregoing teaching content, many modification forms and variation forms are possible. Embodiments are selected and described to fully illustrate the principles of the technical solutions and practical application of the principles, so that other persons skilled in the art can make full use of technical solutions and various embodiments that have various modifications applicable to conceived specific usage.
Claims
1. An electronic device comprising:
- a first part;
- a second part;
- a first display disposed on the first part;
- a second display disposed on the second part;
- a touch panel disposed on the second part;
- one or more processors configured to cause the electronic device to: display an icon of a first application on a home screen on the first display; detect a first gesture on the icon; display, in response to the first gesture, a first graphical user interface (GUI) of the first application on the second display; display, in response to the first gesture, an iconic control on the first display, wherein the iconic control is configured to control the first GUI; detect a second gesture on the touch panel; and display, in response to the second gesture and while displaying the iconic control on the first display, a second GUI of the first application on the second display.
2. The electronic device of claim 1, wherein the one or more processors are further configured to cause the electronic device to:
- detect, in the second gesture, coordinate position values of three touch points at time points;
- determine a speed and a direction of the second gesture based on the coordinate position values and the time points;
- perform, based on the speed and the direction, a fast page flip operation on content of the first application; and
- display the fast page flip operation on the second display.
3. The electronic device of claim 1, wherein the first application is a preset application.
4. The electronic device of claim 1, wherein the second GUI is based on a speed and a direction of the second gesture.
5. The electronic device of claim 1, wherein the second GUI is a catalog of the first application.
6. The electronic device of claim 1, wherein the second gesture comprises touching the touch panel with fingers, keeping the fingers in contact with the touch panel, and moving the fingers away from contact with the touch panel.
7. The electronic device of claim 1, wherein the touch panel is disposed on the second display at a position of the second part.
8. The electronic device of claim 1, wherein the first display is a liquid-crystal display (LCD), and wherein the second display is an electronic ink display.
9. A method implemented by an electronic device and comprising:
- displaying an icon of a first application on a home screen on a first display of the electronic device;
- detecting a first gesture on the icon;
- displaying, in response to the first gesture, a first graphical user interface (GUI) of the first application on a second display of the electronic device;
- displaying, in response to the first gesture, an iconic control on the first display, wherein the iconic control is configured to control the first GUI;
- detecting a second gesture on a touch panel which is disposed on the second display; and
- displaying, in response to the second gesture and while displaying the iconic control on the first display, a second GUI of the first application on the second display.
10. The method of claim 9, further comprising:
- detecting, in the second gesture, coordinate position values of three touch points at time points;
- determining a speed and a direction of the second gesture based on coordinate values of the three touch points and the time points;
- performing, based on the speed and the direction, a fast page flip operation on content of the first application; and
- displaying the fast page flip operation on the second display.
11. The method of claim 9, wherein the first application is a preset application.
12. The method of claim 9, wherein the second GUI is based on a speed and a direction of the second gesture.
13. The method of claim 9, wherein the second GUI is a catalog of the first application.
14. The method of claim 9, wherein the second gesture comprises touching the touch panel with fingers, keeping the fingers in contact with the touch panel, and moving the fingers away from contact with the touch panel.
15. A computer program product comprising instructions that are stored on a computer-readable medium and that, when executed by a processor, cause an electronic device to:
- display an icon of a first application on a home screen on a first display of the electronic device;
- detect a first gesture on the icon;
- display, in response to the first gesture, a first graphical user interface (GUI) of the first application on a second display of the electronic device;
- display, in response to the first gesture, an iconic control on the first display, wherein the iconic control is configured to control the first GUI;
- detect a second gesture on a touch panel which is disposed on the second display; and
- display, in response to the second gesture and while displaying the iconic control on the first display, a second GUI of the first application on the second display.
16. The computer program product of claim 15, wherein the instructions, when executed by the processor, further cause the electronic device to:
- detect, in the second gesture, coordinate position values of three touch points at time points;
- determine a speed and a direction of the second gesture based on coordinate values of the three touch points and the time points;
- perform, based on the speed and the direction, a fast page flip operation on content of the first application; and
- display the fast page flip operation on the second display.
17. The computer program product of claim 15, wherein the first application is a preset application.
18. The computer program product of claim 15, wherein the second GUI is based on a speed and a direction of the second gesture.
19. The computer program product of claim 15, wherein the second GUI is a catalog of the first application.
20. The computer program product of claim 15, wherein the second gesture comprises touching the touch panel with fingers, keeping the fingers in contact with the touch panel, and moving the fingers away from contact with the touch panel.
Type: Application
Filed: Jan 4, 2023
Publication Date: Jan 11, 2024
Inventors: Juan Lan (Shenzhen), Shuangxiong Wen (Shenzhen), Qiang Peng (Shenzhen)
Application Number: 18/149,838