Im investigating using construct to develop a simple real-time instrument music app. However Ive noticed that touch input lag can vary wildly between devices from minimum estimated 40/50ms on newer higher end phones to, on some older or cheaper phones, more than maybe 200ms (making a real-time music app unfeasible for many devices).
I still may go ahead as 40/50ms is getting into the realm of usable but is there anything I can do to reduce it further?
For example, I am assuming "on touch object" or "on any touch start" is the best way to go rather than "on tap" (as this may be waiting for a second tap.)
Are there any other things I should be considering. Is there any code I can add in that optimizes touch input processing etc.
Can stuff like mentioned in this article be of any benefit?
any advice is appreciated cheers.
edit. sorry I just realized Im am being a little dumb as the lag could also be on the audio processor end or other software hardware. Essentially, I am looking to reduce the time as much as possible between touch event and a play sound/ change sound action.
in most games even a 200ms delay may not be noticeable but for a music/sound app any delay is very perceivable and longer delays like 100ms + make real-time playing almost impossible. Im told that Ios processes sound much faster than android but I have not tested any ios devices.
anyone have experience in this area?