Forget about analogies, metaphors, and similes involving light bulbs, coffee machines, operating systems, DAW applications, VSTi virtual instruments, VST effect plug-ins, and all that stuff.
I explored the ACE Studio website; and it's similar to what ElevenLabs provides.
I use ElevenLabs AI voices for some of the characters in my ongoing series of old-time science fiction radio plays; so I have experience using this stuff.
ElevenLabs and ACE Studio are cloud-based services and require an Internet connection for the services they provide.
In the case of ElevenLabs, I have a monthly subscription which costs $22 a month; is billed in advance month-by-month; and does
not require a contract.
If I need a voice for a female character, then I compose a script for what I want her to say and then upload it to ElevenLabs.
ElevenLabs then generates the voice; and I download the MP3/4 audio file of the generated voice.
I studied the information at the ACE Studio website; and it's very similar to the way ElevenLabs works; but ACE Studio is focused more on generating music than generating voices, although ElevenLabs is in beta testing of its own AI music generating technology.
FINANCIAL ASPECTS
The ACE Studio "Artist Pro" subscription costs $22 a month, which is the same as ElevenLabs for its "Creator" subscription; but ElevenLabs also has a "PRO" subscription for $99 per month; and ElevenLabs has Business subscriptions, which are more expensive but provide more minutes and other features.
For what I am doing, the "Creator" subscription is all I need.
ACE Studio uses a two-year "rent-to-own" plan, where after subscribing and paying for two years, you are granted a license to use the cloud-based service.
At $22 per month, this is $264 for each of the two years for a total of $528 (USD).
My perspective based on subscribing and using ElevenLabs AI voice services is that it's worth $22 a month; because it creates male and female voices that I can use to enhance the voice-overs I do myself. If I need a female voice for spoken word in one of the chapters of my old-time science fiction radio play, then I this is what ElevenLabs provides based on the voice actor I select and the script I provide, where the script is simple text with relevant punctuation, for example "Look at the monitor and chart the trajectory of the asteroid."
Then, I edit the audio clip in Fender Studio Pro 8 (FSP8) to adjust the timing and volume level; and I process the edited audio clip with a few VST effect plug-ins, which can include processing with Melodyne to change the vocal pitch and other characteristics, all of which takes the raw audio generated by ElevenLabs and tailors it to what I desire.
THOUGHTS AND SPECIFICS
In all these new AI services, I think it makes sense to focus on specific details and the way things happen.
For ACE Studio, the marketing blurb is that everything is done in what I call an
automagical way which does not use traditional VSTi virtual instruments and sampled sound libraries and VST effect plug-ins.
However, ACE Studio partnered with EastWest in January 2026; and since I have been using EastWest instruments and sounds for years, I know what EastWest does and what EastWest can provide, which basically is technology and sampled sounds for virtual instruments and virtual voices.
ACE Studio Partners with EASTWEST Sounds
This is fine with me; and it looks to provide ACE Studio with oodles of sampled-sound libraries, since at present EastWest has over 40,000 virtual instruments and sounds.
Regarding using ACE Studio, I watched the YouTube video; and the process begins by the user providing MIDI sequences and various information to tell ACE Studio what the user desires to happen.
In this respect, ACE Studio is
not psychic, hence needs information from the user to develop an accurate sense of what the user desires.
This ACE Studio web page provides a short, high-level overview of the AI Violin and how to use it; but it does not explain how to provide the information necessary for ACE Studio to determine how to do the desired articulations, playing styles, dynamics, and all that detailed music notation, VSTi virtual instrument, and VST effect plug-in stuff.
AI Violin (ACE Studio)
Yet, although I do not have a subscription for ACE Studio, I am certain the user needs to provide more information than just "I like violins".
As noted, neither ACE Studio nor ElevenLabs is psychic.
If anyone needs a clue, then ask Amazon Alexa, Apple Siri, Google AI, or ChatGPT, "What am I thinking?"
ChatGPT
Based on the information on the AI Violin at the ACE Studio website, it needs to be provided with a MIDI sequence to generate the violin audio; and in the same or similar way as Tonalic (Celemony), I think it needs more information to know what to do.
SUMMARY
If it's possible to try ACE Studio for one-month only at $22
without being obligated to continue the service and
without any cancellation fees or penalties, then I might try it to develop a perspective on how it works and what it does.
I do everything with music notation in FSP8; but as demonstrated in the YouTube video I made earlier this morning (Sunday February 8, 2026) to help a forum member learn how to use the FSP8 Arranger Track, Signature Track, and Tempo Track to specify various time signatures and tempos for the sound track of a film, it's easy in FSP8 to switch from (a) music notation to (b) MIDI and vice-versa; hence the ACE Studio requirement for MIDI is not a problem for me.
Since at minimum, I have to provide MIDI or music notation, I am not certain what advantage using ACE Studio can provide, especially since I do
not specify articulations, playing styles, dynamics, and all that visually cluttering music notation nonsense, preferring instead (a) to use sampled-sound libraries where the trained musicians are playing in the desired articulations and playing styles, (b) to use VST effect plug-ins to control dynamics, and (c) for what I call "diatonically" sampled-sound libraries where only every other note is sampled, which requires the in-between notes to be computer-generated using logarithmic interpolation, I use dry samples and then add motion effects like tremolo, vibrato, and echoes via VST effect plugins, FSP8 Automation, Melodyne (Celemony) and whatever else avoids having motion embedded in the sampled sounds, which is based on the rule that it's easy to add stuff to audio but nearly impossible to remove when it's embedded in the audio.
Lots of FUN
P. S. If this is too much information and too many words, then it's not my fault. I do this every day somewhere and have been writing tens of thousands of words every day for decades. I enjoy touch-typing and composing prose in real-time on the fly. It's one of the things I do; and I hope it helps folks, but regardless (a) it helps me and (b) it's a key aspect of the way I learn new stuff.
I hope the information in this post is helpful!