Connect with us

Technology

AI Technology Makes It Possible to 3D Scan and Edit Real-World Objects

Published

on

Imagine being able to view a realistic, fully editable 3D model of an object from any angle by simply sweeping your smartphone around it. AI advancements are making this a reality very quickly.

For precisely this purpose, Canadian researchers at Simon Fraser University (SFU) have unveiled new AI technology. Consumers will soon be able to take 3D captures of real-life objects instead of just 2D ones, and edit their shapes and appearance as they like, with the same ease as they can do with regular 2D photos.

Researchers presented Proximity Attention Point Rendering (PAPR), a novel method that converts a collection of 2D images of an object into a cloud of 3D points that depicts the object’s shape and appearance, at the 2023 Conference on Neural Information Processing Systems (NeurIPS) in New Orleans, Louisiana. The paper was published on the arXiv preprint server.

After that, each point has a knob that allows users to manipulate the object: dragging a point modifies its shape, and editing its properties alters its appearance. Afterwards, through a procedure called “rendering,” the 3D point cloud can be viewed from any perspective and converted into a 2D image that accurately depicts the edited object from the perspective from which the image was taken.

Researchers demonstrated how to bring a statue to life using the new AI technology. The technology automatically turned a collection of images of the statue into a 3D point cloud, which is subsequently animated. The final product is a video that shows the statue moving its head side to side while a path is shown around it.

A paradigm shift in the reconstruction of 3D objects from 2D images is primarily being driven by AI and machine learning. According to Dr. Ke Li, senior author of the paper and assistant professor of computer science at Simon Fraser University (SFU), “The outstanding success of machine learning in fields like computer vision and natural language is inspiring researchers to investigate how traditional 3D graphics pipelines can be re-engineered with the same deep learning-based building blocks that were responsible for the recent wave of AI success stories.”

“It turns out that doing so successfully is a lot harder than we anticipated and requires overcoming several technical challenges. What excites me the most is the many possibilities this brings for consumer technology—3D may become as common a medium for visual communication and expression as 2D is today.”

Creating a 3D representation of shapes that is easy and intuitive for users to edit is one of the main challenges in 3D modeling. Neural radiance fields (NeRFs) are one prior method that requires the user to describe what happens to each continuous coordinate, making shape editing difficult. The shape surface may be crushed or shattered after editing, which makes a more modern method called 3D Gaussian splatting (3DGS) unsuitable for shape editing as well.

One of the researchers’ most important insights was realizing that every 3D point in the point cloud could be thought of as a control point in a continuous interpolator rather than as a discrete splat. Subsequently, the shape automatically and intuitively changes when the point is moved. In animated videos, animators use a similar method to define object motion: they specify an object’s position at a few points in time, and an interpolator automatically generates the object’s motion at every point in time.

Nevertheless, it is not easy to define an interpolator mathematically between any random set of 3D points. The scientists developed a machine learning model that uses proximity attention, a novel mechanism, to learn the interpolator in an end-to-end manner.

A spotlight at the NeurIPS conference, an honor given to the top 3.6% of paper submissions, was given to the paper in appreciation of this technological advancement.

The research team is looking forward to the future with excitement. “This opens the way to many applications beyond what we’ve demonstrated,” Dr. Li added. “We are already exploring various ways to leverage PAPR to model moving 3D scenes and the results so far are incredibly promising.”

Ke Li, Yanshu Zhang, Shichong Peng, and Alireza Moazeni are the paper’s authors. Zhang, Peng, and Moazeni are Ph.D. candidates in the School of Computing Science at Simon Fraser University (SFU), and Zhang and Peng are co-first authors.

Technology

Microsoft Expands Copilot Voice and Think Deeper

Published

on

Microsoft Expands Copilot Voice and Think Deeper

Microsoft is taking a major step forward by offering unlimited access to Copilot Voice and Think Deeper, marking two years since the AI-powered Copilot was first integrated into Bing search. This update comes shortly after the tech giant revamped its Copilot Pro subscription and bundled advanced AI features into Microsoft 365.

What’s Changing?

Microsoft remains committed to its $20 per month Copilot Pro plan, ensuring that subscribers continue to enjoy premium benefits. According to the company, Copilot Pro users will receive:

  • Preferred access to the latest AI models during peak hours.
  • Early access to experimental AI features, with more updates expected soon.
  • Extended use of Copilot within popular Microsoft 365 apps like Word, Excel, and PowerPoint.

The Impact on Users

This move signals Microsoft’s dedication to enhancing AI-driven productivity tools. By expanding access to Copilot’s powerful features, users can expect improved efficiency, smarter assistance, and seamless integration across Microsoft’s ecosystem.

As AI technology continues to evolve, Microsoft is positioning itself at the forefront of innovation, ensuring both casual users and professionals can leverage the best AI tools available.

Stay tuned for further updates as Microsoft rolls out more enhancements to its AI offerings.

Continue Reading

Technology

Google Launches Free AI Coding Tool for Individual Developers

Published

on

Google Launches Free AI Coding Tool for Individual Developers

Google has introduced a free version of Gemini Code Assistant, its AI-powered coding assistant, for solo developers worldwide. The tool, previously available only to enterprise users, is now in public preview, making advanced AI-assisted coding accessible to students, freelancers, hobbyists, and startups.

More Features, Fewer Limits

Unlike competing tools such as GitHub Copilot, which limits free users to 2,000 code completions per month, Google is offering up to 180,000 code completions—a significantly higher cap designed to accommodate even the most active developers.

“Now anyone can easily learn, generate code snippets, debug, and modify applications without switching between multiple windows,” said Ryan J. Salva, Google’s senior director of product management.

AI-Powered Coding Assistance

Gemini Code Assist for individuals is powered by Google’s Gemini 2.0 AI model and offers:
Auto-completion of code while typing
Generation of entire code blocks based on prompts
Debugging assistance via an interactive chatbot

The tool integrates with popular developer environments like Visual Studio Code, GitHub, and JetBrains, supporting a wide range of programming languages. Developers can use natural language prompts, such as:
Create an HTML form with fields for name, email, and message, plus a submit button.”

With support for 38 programming languages and a 128,000-token memory for processing complex prompts, Gemini Code Assist provides a robust AI-driven coding experience.

Enterprise Features Still Require a Subscription

While the free tier is generous, advanced features like productivity analytics, Google Cloud integrations, and custom AI tuning remain exclusive to paid Standard and Enterprise plans.

With this move, Google aims to compete more aggressively in the AI coding assistant market, offering developers a powerful and unrestricted alternative to existing tools.

Continue Reading

Technology

Elon Musk Unveils Grok-3: A Game-Changing AI Chatbot to Rival ChatGPT

Published

on

Elon Musk Unveils Grok-3: A Game-Changing AI Chatbot to Rival ChatGPT

Elon Musk’s artificial intelligence company xAI has unveiled its latest chatbot, Grok-3, which aims to compete with leading AI models such as OpenAI’s ChatGPT and China’s DeepSeek. Grok-3 is now available to Premium+ subscribers on Musk’s social media platform x (formerly Twitter) and is also available through xAI’s mobile app and the new SuperGrok subscription tier on Grok.com.

Advanced capabilities and performance

Grok-3 has ten times the computing power of its predecessor, Grok-2. Initial tests show that Grok-3 outperforms models from OpenAI, Google, and DeepSeek, particularly in areas such as math, science, and coding. The chatbot features advanced reasoning features capable of decomposing complex questions into manageable tasks. Users can interact with Grok-3 in two different ways: “Think,” which performs step-by-step reasoning, and “Big Brain,” which is designed for more difficult tasks.

Strategic Investments and Infrastructure

To support the development of Grok-3, xAI has made major investments in its supercomputer cluster, Colossus, which is currently the largest globally. This infrastructure underscores the company’s commitment to advancing AI technology and maintaining a competitive edge in the industry.

New Offerings and Future Plans

Along with Grok-3, xAI has also introduced a logic-based chatbot called DeepSearch, designed to enhance research, brainstorming, and data analysis tasks. This tool aims to provide users with more insightful and relevant information. Looking to the future, xAI plans to release Grok-2 as an open-source model, encouraging community participation and further development. Additionally, upcoming improvements for Grok-3 include a synthesized voice feature, which aims to improve user interaction and accessibility.

Market position and competition

The launch of Grok-3 positions xAI as a major competitor in the AI ​​chatbot market, directly challenging established models from OpenAI and emerging competitors such as DeepSeek. While Grok-3’s performance claims are yet to be independently verified, early indications suggest it could have a significant impact on the AI ​​landscape. xAI is actively seeking $10 billion in investment from major companies, demonstrating its strong belief in their technological advancements and market potential.

Continue Reading

Trending

error: Content is protected !!