Connect with us

Technology

Launched by Visual Electric to free AI art creation from chat interfaces

Published

on

Launched by Visual Electric to free AI art creation from chat interfaces

There are probably some similarities that you have observed if you have experimented with at least a few of the text-to-image AI art generation services that have been introduced in recent years, like Midjourney or OpenAI’s different DALL-E versions. The most notable of them all is that they all resemble chat interfaces. The application usually responds with an image embedded in a message after the user enters their text prompts.

While this point of interaction functions admirably for some clients and application engineers, certain individuals accept it is restricting and at last not what laid out craftsmen and architects need while utilizing computer based intelligence at work. However, presently San Francisco-based Visual Electric is here to offer an alternate methodology. One that the new startup — which rises up out of covertness today following a seed round last year from Sequoia, BoxGroup, and Creator Asset of an undisclosed total — accepts is preferable adjusted to visual imagination over messaging to and fro with a man-made intelligence model.

“There’s just so many workflow-specific optimizations that you need to make if you’re a graphic designer or a concept artist,” said Colin Dunn, founder and CEO of Visual Electric, in an exclusive interview with VentureBeat. “There’s a long tail of things that will make their life way easier and will make for a much better product.”

Dunn recently drove item plan and brand at the versatile site building organization Universe, and before that, filled in as head of plan at Playspace, a Google procurement.

For enterprise users, such as independent designers, in-house designers at major brands, and even “pro-sumers,” Visual Electric aims to be that “much better product” for AI art, visual design, and creativity.

The organization is intentionally not sending off its own hidden artificial intelligence picture generator AI (ML) model. Instead, it is based on the open-source Stable Diffusion XL model, which is currently the subject of a copyright lawsuit brought by artists against Stability AI, the company that developed it, as well as Midjourney and other AI art generators.

This is due to the fact that Dunn and his two co-founders, Adam Menges, chief product officer of Visual Electric and former co-founder of Microsoft acquisition Lobe; and chief technology officer Zach Stiggelbout, who was previously employed by Lobe, are of the opinion that image generation AI models are in the process of being commoditized, and that the front-end user interface will largely determine the success and failure of businesses.

“We just want to build the best product experience,” Dunn said. “We’re really model agnostic and we’re happy to swap out whatever model is going to give users the best results. Our product can easily accommodate multiple models or the next model that’s going to come out.”

What sets Visual Electric apart from Midjourney, DALL-E 3, and other AI art apps?

What sets Visual Electric apart from previous image generators? Instead of the top-to-bottom “linear” form factor of other chat-based AI art generator apps, which force users to scroll back up to see their previous generations, it allows users to generate and drag-to-move their imagery around an infinite virtual “canvas.” Clients can keep producing new arrangements of 4 pictures all at once and move them around this material any place they’d like.

“Creativity is a nonlinear process,” Dunn said. “You want to explore; you want to go down different paths and then go back up to an idea you were looking at previously and take that in a new direction. Chat forces you into this very linear flow where it’s sort of like you have a starting point and an ending point. And that’s just not really how creativity works.”

Unlike many chat interfaces, this box has been moved to the top of the screen instead of the bottom, although there is still a space for text prompts to be entered.

To assist with conquering the underlying obstacle that a few clients face — not knowing precisely exact thing to type in to provoke the computer based intelligence to inspire it to create the picture they have to their eye — Visual Electric offers a drop-down field of autocomplete ideas, like what a client finds while composing in a pursuit on Google. All of these suggestions are based on what Visual Electric has observed from early users and what produces the best images. In any case, a client is likewise allowed to veer off from these completely and type in a custom brief too.

Moreover, Visual Electric’s electronic man-made intelligence workmanship generator offers a scope of supportive extra devices for changing the brief and style of the subsequent pictures, remembering pre-set styles that emulate normal ones for the pre-man-made intelligence computerized and printed craftsmanship universes, including “marker,” “exemplary movement,” “3D render,” “digitally embellish,” “risograph,” “stained glass,” and numerous others — with recent trends being added routinely.

It puts it in more direct competition with Adobe’s Firefly 2 AI art interface, which offers similar functionality, as the user can select their image aspect ratio from buttons on the dropdown or a convenient right-rail sidebar rather than having to specify it within the prompt text. Two common examples of this are 16:9 and 5:4.

This sidebar additionally allows the client to determine prevailing varieties and components they wish to reject from their subsequent simulated intelligence created picture, likewise inputted through text.

In addition, the user can click a button to “remix” or “regenerate” their images based on their initial prompt, or they can “touch up” specific areas of the image and have the AI regenerate only those areas that they highlight using a digital brush of a size that the user can adjust, while keeping the rest of the image intact and adding to it in the same way. So, for instance, you could “touch up” the hair of your AI-generated subject and instruct the Stable Diffusion XL model to redo only that portion of the image if you didn’t like it.

Additionally, there is a built-in upscaler that can improve image resolution and detail.

“These are the tools that represent what we see as the AI-native workflow and they in the order that you use them,” Dunn said.

Pricing, the community, and early success stories

Despite the fact that Visual Electric is going public today, the company has been quietly conducting alpha testing with a few dozen designers. Dunn claims that these designers have already provided valuable feedback that will help improve the product. Additionally, Dunn says that the promising results of how Visual Electric has been used to assist in real-world enterprise workplace situations show that the company is on the right track.

Dunn referenced one client specifically — keeping the name for classification — who had a little group of creators attempting to make menus and other visual guarantee for in excess of 600 colleges.

Previously, this group would have invested bunches of their energy figuring out stock symbolism and trying to track down pictures that matched each other yet likewise addressed genuinely the things on a school’s eating corridor menu, and having to physically alter the stock symbolism to make it more precise.

With Visual Electric, they can now create brand-new images that meet the requirements of the menu and edit portions of them without using Adobe Photoshop or other alternatives.

“They’re now able to take what was a non-creative task and make it into something that is very creative, much more fulfilling, and they can do it in a tenth of the time,” Dunn claimed.

An “Inspiration” feed of AI-generated images created on the platform by other users is another important feature that Visual Electric offers. This feed, a lattice of various estimated pictures that inspires Pinterest, permits the client to float over the pictures and see their prompts. They can also import any images from the public feed into their private canvas by “remixing” them.

“This was a early decision that we made, which is we think that with generative AI there’s an opportunity to bring the network into the tool,” Dunn explained. “Right now, you have inspiration sites like Pinterest and designer-specific sites like Dribbble, and then you have the tools like Photoshop, Creative Suite and Figma. It’s always felt odd to me that these things are not unified in some way, because they’re so related to each other.”

Clients of Visual Electric can decide to draw in with this feed and add to it or not, at their tact. For undertakings worried about the security of their symbolism and works underway, Dunn guaranteed VentureBeat that the organization views security and security in a serious way, however just the “Genius” plan offers the capacity to have secretly put away pictures — all the other things is public as a matter of course.

Sending off in the U.S. today freely, Visual Electric’s valuing is as per the following: a free plan that gives you 40 generations per day at slower speeds and a license that can only be used for personal use (you can’t sell the images or use them for marketing); a standard arrangement at $20 each month or $16/month paid every year direct, which takes into consideration local area sharing, limitless ages at 2x quicker velocities, and sovereignty free business use permit; as well as a well conceived plan for $60 each month or $48/month paid yearly direct, which offers all that the last two plans offer yet additionally significantly higher goal pictures, and fundamentally, privatized ages.

Continue Reading
Advertisement

Technology

Google I/O 2024: Top 5 Expected Announcements Include Pixie AI Assistant and Android 15

Published

on

The largest software event of the year for the manufacturer of Android, Google I/O 2024, gets underway in Mountain View, California, today. The event will be livestreamed by the corporation starting at 10:00 am Pacific Time or 10:30 pm Indian Time, in addition to an in-person gathering at the Shoreline Amphitheatre.

During the I/O 2024 event, Google is anticipated to reveal a number of significant updates, such as details regarding the release date of Android 15, new AI capabilities, the most recent iterations of Wear OS, Android TV, and Google TV, as well as a new Pixie AI assistant.

Google I/O 2024’s top 5 anticipated announcements are:

1) The Android 15 is Highlighted:

It is anticipated that Google will reveal a sneak peek at the upcoming Android version at the I/O event, as it does every year. Google has arranged a meeting to go over the main features of Android 15, and during the same briefing, the tech giant might possibly disclose the operating system’s release date.

While a significant design makeover isn’t anticipated for Android 15, there may be a number of improvements that will assist increase user productivity, security, and privacy. A number of other new features found in Google’s most recent operating system include partial screen sharing, satellite connectivity, audio sharing, notification cooldown, app archiving, and notification cooldown.

2) Pixie AI Assistant:

Also anticipated from Google is the introduction of “Pixie,” a brand-new virtual assistant that is only available on Pixel devices and is powered by Gemini. In addition to text and speech input, the new assistant might also allow users to exchange images with Pixie. This is known as multimodal functionality.

Pixie AI may be able to access data from a user’s device, including Gmail or Maps, according to a report from the previous year, making it a more customized variant of Google Assistant.

3) Gemini AI Upgrades:

The highlight of Google’s I/O event last year was AI, and this year, with OpenAI announcing its newest large language model, GPT-4, just one day before I/O 2024, the firm faces even more competition.

With the aid of Gemini AI, Google is anticipated to deliver significant enhancements to a number of its primary programs, including Maps, Chrome, Gmail, and Google Workspace. Furthermore, Google might be prepared to use Gemini in place of Google Assistant on all Android devices at last. The Gemini AI app already gives users the option to switch the chatbot out as Android’s default assistant app.

4) Hardware Updates:

Google has been utilizing I/O to showcase some of its newest devices even though it’s not really a hardware-focused event. For instance, during the I/O 2023 event, the firm debuted the Google Pixel 7a and the first-ever Pixel Fold.

But, considering that it has already announced the Pixel 8a smartphone, it is unlikely that Google would make any significant hardware announcements this time around. The Pixel Fold series, on the other hand, might be introduced this year alongside the Pixel 9 series.

5) Wear OS 5:

At last, Google has made the decision to update its wearable operating system. But the business has a history of keeping quiet about all the new features that Wear OS 5 will.

A description of the Wear OS5 session states that the new operating system will include advances in the Watch Face format, along with how to build and design for an increasing range of devices.

Continue Reading

Technology

A Vision-to-Language AI Model Is Released by the Technology Innovation Institute

Published

on

The large language model (LLM) has undergone another iteration, according to the Technology Innovation Institute (TII) located in the United Arab Emirates (UAE).

An image-to-text model of the new Falcon 2 is available, according to a press release issued by the TII on Monday, May 13.

Per the publication, the Falcon 2 11B VLM, one of the two new LLM versions, can translate visual inputs into written outputs thanks to its vision-to-language model (VLM) capabilities.

According to the announcement, aiding people with visual impairments, document management, digital archiving, and context indexing are among potential uses for the VLM capabilities.

A “more efficient and accessible LLM” is the goal of the other new version, Falcon 2 11B, according to the press statement. It performs on par with or better than AI models in its class among pre-trained models, having been trained on 5.5 trillion tokens having 11 billion parameters.

As stated in the announcement, both models are bilingual and can do duties in English, French, Spanish, German, Portuguese, and several other languages. Both provide unfettered access for developers worldwide as they are open-source.

Both can be integrated into laptops and other devices because they can run on a single graphics processing unit (GPU), according to the announcement.

The AI Cross-Center Unit of TII’s executive director and acting chief researcher, Dr. Hakim Hacid, stated in the release that “AI is continually evolving, and developers are recognizing the myriad benefits of smaller, more efficient models.” These models offer increased flexibility and smoothly integrate into edge AI infrastructure, the next big trend in developing technologies, in addition to meeting sustainability criteria and requiring less computer resources.

Businesses can now more easily utilize AI thanks to a trend toward the development of smaller, more affordable AI models.

“Smaller LLMs offer users more control compared to large language models like ChatGPT or Anthropic’s Claude, making them more desirable in many instances,” Brian Peterson, co-founder and chief technology officer of Dialpad, a cloud-based, AI-powered platform, told PYMNTS in an interview posted in March. “They’re able to filter through a smaller subset of data, making them faster, more affordable, and, if you have your own data, far more customizable and even more accurate.”

Continue Reading

Technology

European Launch of Anthropic’s AI Assistant Claude

Published

on

Claude, an AI assistant, has been released in Europe by artificial intelligence (AI) startup Anthropic.

Europe now has access to the web-based Claude.ai version, the Claude iOS app, and the subscription-based Claude Team plan, which gives enterprises access to the Claude 3 model family, the company announced in a press statement.

According to the release, “these products complement the Claude API, which was introduced in Europe earlier this year and enables programmers to incorporate Anthropic’s AI models into their own software, websites, or other services.”

According to Anthropic’s press release, “Claude has strong comprehension and fluency in French, German, Spanish, Italian, and other European languages, allowing users to converse with Claude in multiple languages.” “Anyone can easily incorporate our cutting-edge AI models into their workflows thanks to Claude’s intuitive, user-friendly interface.”

The European Union (EU) has the world’s most comprehensive regulation of AI , Bloomberg reported Monday (May 13).

According to the report, OpenAI’s ChatGPT is receiving privacy complaints in the EU, and Google does not currently sell its Gemini program there.

According to the report, Anthropic’s CEO, Dario Amodei, told Bloomberg that the company’s cloud computing partners, Amazon and Google, will assist it in adhering to EU standards. Additionally, Anthropic’s software is currently being utilized throughout the continent in the financial and hospitality industries.

In contrast to China and the United States, Europe has a distinct approach to AI that is characterized by tighter regulation and a stronger focus on ethics, PYMNTS said on May 2.

While the region has been sluggish to adopt AI in vital fields like government and healthcare, certain businesses are leading the way with AI initiatives there.

In numerous areas, industry benchmark evaluations of Anthropic’s Claude 3 models—which were introduced in 159 countries in March—bested those of rival AI models.

On May 1, the business released its first enterprise subscription plan for the Claude chatbot along with its first smartphone app.

The introduction of these new products was a major move for Anthropic and put it in a position to take on larger players in the AI space more directly, such as OpenAI and Google.

Continue Reading

Trending

error: Content is protected !!