Artificial Intelligence News - Page 3

The latest and most important Artificial Intelligence news - Page 3.

'Superhuman' AI beats pro players & hits Grandmaster Starcraft II rank

Jak Connor | Fri, Nov 1 2019 4:13 AM CDT

An artificial intelligence developed by Google's AI firm DeepMind has managed to achieve the highest level of ranking in Starcraft II.

According to the results which were published in Nature, DeepMind's AI was released onto the European Starcraft II servers and placed within the top 0.15% of the regions 90,000 players. Jon Dodge, an AI researcher at Oregon State University in Corvallis was shocked at the progress the AI made, saying "I did not expect AI to essentially be superhuman in this domain so quickly, maybe not for another couple of years."

The AI is called AlphaStar and before it was released onto the European servers its speed was reduced to make a fairer contests. The researchers wanted players to also not know they were versing an AI, David Silver, who co-leads the AlphaStar project said "We wanted this to be like a blind experiment. We really wanted to play under those conditions and really get a sense of, 'how well does this pool of humans perform against us?'"

Continue reading: 'Superhuman' AI beats pro players & hits Grandmaster Starcraft II rank (full post)

Deep neural network achieves human-like character movement & motion

Jak Connor | Wed, Oct 30 2019 1:08 AM CDT

Researchers have showcased a brand new 3D animated character being brought to life, the catch? It moves exactly like humans do.

Deep neural network achieves human-like character movement & motion | TweakTown.com

Computer scientists from the University of Edinburgh and Adobe Research have developed a data driven technique that uses a deep neural network to accurately guide 3D animated characters. The precision of the neural network showcases characters in a variety of different motions such as sitting in chairs, picking up objects, running, side-stepping, climbing through obstacles and more.

Komura, coauthor and chair of computer graphics at the University of Edinburgh spoke out about the achievement, saying "The technique essentially mimics how a human intuitively moves through a scene or environment and how it interacts with objects, realistically and precisely." If you are interested in checking out the video of the animated character for yourself, click this link right here.

Continue reading: Deep neural network achieves human-like character movement & motion (full post)

Facebook's new AI system can 'de-identify' peoples faces in real-time

Jak Connor | Sun, Oct 27 2019 1:03 AM CDT

Facebook researchers have announced the development of a new AI system that will assist in de-identifying peoples faces.

Facebook's new AI system can 'de-identify' peoples faces in real-time | TweakTown.com

While you would originally think that Facebook would be working on an facial recognition software that is designed to identify peoples faces, that's exactly the opposite of what they doing. A new announcement out of Facebook Research shines a light on a new AI system that is designed to ever so slightly distort peoples faces.

Above you can see an example of Jennifer Lawrence's face being distorted. From the image, you can see what the AI generated isn't that different when compared to the original, but is different enough that facial recognition technology would have a much harder time identifying its her. This new technology is designed to assist people in keeping their identification secure from third-party facial recognition software that could potentially scam users. For more information about the new AI, visit the Facebook Research website here.

Continue reading: Facebook's new AI system can 'de-identify' peoples faces in real-time (full post)

This AI can spot brain hemorrhages with INSANE pixel-level accuracy

Jak Connor | Thu, Oct 24 2019 5:30 AM CDT

Since their discovery, brain hemorrhages have always been a doctors nightmare as missing even the tiniest hemorrhage can leave the patient in a fatal state. Now the responsibility might not all be on the doctors, AI has now been developed to shoulder some as well.

This AI can spot brain hemorrhages with INSANE pixel-level accuracy 02 | TweakTown.com

UC Berkeley and UCSF researchers have managed to conceive an algorithm that is able to detect brain hemorrhages with better accuracy than two out of four radiologists. This algorithm was created using massive amounts of data, 4,396 CT scans and convolutional neural network. While that sample size of CT scans might sound relatively on the smaller side, it should be noted that the AI was able to detect abnormalities within the scans "at the pixel level".

This means that the AI is able to decipher noise and other errors, that normal human doctors may run into, out of the equation. Therefore giving a more technical analysis on the brain scans which would then result in a more accurate assessment of what needs to be done. While you might think that is AI will be replacing doctors, it won't be. Instead, it will be assisting doctors in discovering abnormalities that they might of originally missed, essentially saving the doctors massive amounts of time.

Continue reading: This AI can spot brain hemorrhages with INSANE pixel-level accuracy (full post)

Deep learning AI beats expert scholars at deciphering ancient texts

Jak Connor | Mon, Oct 21 2019 6:09 AM CDT

Researchers at the University of Oxford built and trained a neural network to be able to fill in the letter gaps of ancient texts, and now the AI is better than expert scholars.

Deep learning AI beats expert scholars at deciphering ancient texts | TweakTown.com

The researchers tested the AI on ancient Greek inscriptions that were on objects such as stones, ceramics and metal. The texts are dated back to 1500 and 2600 years ago, and according to a report out of New Scientist, the AI creamed the humans in a head-to-head speed test at deciphering the artifacts. "In a head-to-head test, where the AI attempted to fill the gaps in 2949 damaged inscriptions, human experts made 30 percent more mistakes than the AI. Whereas the experts took 2 hours to get through 50 inscriptions, Pythia gave its guesses for the entire cohort in seconds."

New Scientist says that the AI which has been titled as Pythia was able to recognize and remember patterns in 35,000 different relics that amassed over 3 million words. It was also able to pick up patterns and include context such as the shape and layout within its descriptions. Pythia gives scholars predictions for missing letters or words within the text and rather than returning to the scholars with a single prediction Pythia gives multiple predictions and its level of confidence for each one.

Continue reading: Deep learning AI beats expert scholars at deciphering ancient texts (full post)

OpenAI shows one-handed robot solve a Rubik's cube

Anthony Garreffa | Tue, Oct 15 2019 11:09 PM CDT

We all know our days are numbered, with our AI and robotic overlords planning to overthrow humanity at some point in the future... and it all seems like it'll begin with a Rubik's Cube.

AI research organization OpenAI have been hard at work building a general purpose, self-learning robot with its robotics division Dactyl unveiling its humanoid robotic hand in 2018 -- which is now being used to solve a Rubik's cube in less than 4 minutes flat. OpenAI is working on a number of different robotic parts with its in-house AI software, with this robotic arm just one of those.

Dactyl stumbles, but eventually solves the Rubik's cube -- with the team having a goal of seeing their AI-powered robotic appendages working on real-world tasks. Their robots packed with AI can learn real-world things, and won't need to be specifically programmed. This means that Dactyl is a self-learning robotic hand that looks at new tasks just like you and I would.

Continue reading: OpenAI shows one-handed robot solve a Rubik's cube (full post)

Bicycles, dinner tables & umbrellas could soon become pocket-sized

Jak Connor | Tue, Oct 15 2019 2:08 AM CDT

Have you ever wondered if it would be possible for you to pick up your bicycle, fold it into itself and then place it in your pocket? Well, this new super-compressible material could do just that.

Researchers at TU Delft have used artificial intelligence to create a new supercompressible but strong material. According to Miguel Bessa, assistant professor in materials science and engineering at TU Delft, the idea originiated when he was at the California Institute of Technology in the corner of the Space Structures Lab. Bessa noticed that a satellite structure could open long solar sails from an extremely small form-factor.

This observation drove Bessa's inspiration to create a supercompressible material that could be compressed into a fraction of its volume, but still remain strong. "If this was possible, everyday objects such as bicycles, dinner tables and umbrellas could be folded into your pocket." Bessa and his team used artificial intelligence instead of the traditional trial-and-error process to explore new design possibilities with metamaterials. This reduced experimentation to the absolute minimum, and after some time Bessa fabricated two designs that converted once brittle polymers into lightweight, recoverable and super-compressible metamaterials.

Continue reading: Bicycles, dinner tables & umbrellas could soon become pocket-sized (full post)

AI doesn't know why people with bigger hands have larger vocabularies

Jak Connor | Mon, Oct 14 2019 2:03 AM CDT

While it might seem like a silly idea at first, did you know that people with large hands actually have bigger vocabularies than people with small hands? Its true.

AI doesn't know why people with bigger hands have larger vocabularies | TweakTown.com

Dr. Gary Marcus, the director of the NYU Infant Language Learning Center, and a professor of psychology at New York University has spoken out about this very topic and how artificial intelligence (AI) is also thrown into the mix. Marcus says this is an old joke that is tossed around by statisticians, and when a person takes into account the entire population and measure everyone's hand-size, the people with larger hands will have larger vocabularies. This is purely because of the fact that the people with larger hands tend to be older, and that adults tend to know more words then children.

This is correlated evidence, and not causation. Saying that something is causing people to learn new words, and causing them to grow their hands at the same time is an observed correlation between the two measured groups. Saying that growing your hand made your vocabulary grow is suggesting causation, this is a very important distinguishable definition that us humans can understand quite easily. Artificial intelligence on the other hand struggles to see the relationship between the two.

Continue reading: AI doesn't know why people with bigger hands have larger vocabularies (full post)

This AI can determine how effective antidepressants are on patients

Jak Connor | Thu, Sep 26 2019 3:07 AM CDT

Two new studies have been published by UT Southwestern and have shown evidence towards an artificial intelligence (AI) being able to determine whether antidepressants work on select patients.

This AI can determine how effective antidepressants are on patients | TweakTown.com

With depression being a rampant disease in this day and age, scientists are working around the clock to counteract its widespread effects. The team out of UT Southwestern have used an AI to identify patterns of brain activity that allow them to determine whether or not an individual patient is responsive to certain antidepressants. This evidence suggests that if scientists can accurately scan patients brain activity they could be able to pinpoint whether or not the medication that has been prescribed to them is effective or not.

The complications that the team underwent is the fact that depression is extremely hard to pinpoint, as it can exist in various states of the brain. This makes it hard to get an accurate representation of how it manifests per patient and requires researchers to scan the brain in different states, eg emotion induced/resting. Dr. Madhukar Trivedi, founding Director of UT Southwestern's Center for Depression Research and Clinical Care said "Depression is a complex disease that affects people in different ways. Much like technology can identify us through fingerprints and facial scans, these studies show we can use imaging to identify specific signatures of depression in people."

Continue reading: This AI can determine how effective antidepressants are on patients (full post)

Google releases 3000 deepfakes to help make deepfake detection tools

Anthony Garreffa | Wed, Sep 25 2019 8:42 AM CDT

We are quickly descending into the insane reality of deepfake technology, where it will be virtually impossible to know the difference between a real or fake video of someone -- and Google wants to help.

This deepfake of Bill Hader impersonating Tom Cruise would have to be my favorite go-to deepfake, with every single person I've showed it to (even those aware of what deepfake is) absolutely blown away by it.

The search giant has released a bunch of deepfake videos that will help researchers make deepfake detection tools, because we're at a stage when the fear factor hits 11/10 and people are scared of them being used in the upcoming 23020 presidential elections.

Google filmed actors in a bunch of different poses and scenes, and then used publicluy available deepfake generation methods to create 3000 new deepfakes. Researchers now have access to this trove of deepfakes, and can use them to train automated detection tools to make them better at detecting real, and not-so-real videos.

Continue reading: Google releases 3000 deepfakes to help make deepfake detection tools (full post)

Newsletter Subscription
Latest News
View More News
Latest Reviews
View More Reviews
Latest Articles
View More Articles