OpenAI Residency

OpenAI Residency

As part of our effort to support and develop AI talent, we’re excited to announce the OpenAI Residency. This new program offers a pathway to a full-time role at OpenAI for researchers and engineers who don’t currently focus on artificial intelligence. We are excited to get applications from everyone, and will make a special effort to hear from underrepresented groups in technology.

The program is an iteration of our former Scholars and Fellows programs. The Residency shifts the focus away from curriculum-based learning, instead giving Residents an opportunity to work collaboratively alongside OpenAI teams on active projects.

The first cohort of the six-month program begins in April 2022 and Residents will be compensated as fully salaried employees for the duration of the program.

“There are many talented people who want to contribute to AI but cannot find an easy way to do so,” said Ilya Sutskever, OpenAI’s Chief Scientist. “The Residency aims to address that, by teaching participants the most important practical AI skills in a hands-on way as quickly as possible. We’ve welcomed incredible new talent to OpenAI through our Fellows and Scholars programs, who have made major research contributions and helped advance OpenAI’s goal of building beneficial AGI.”

Over the last three years we’ve made more than 20 full-time hires through our mentorship programs, representing one in six members of our technical staff, and our new iteration will broaden the range of candidates we are considering.

Excellent work and experience can come from both inside and outside of the traditional education and work settings. OpenAI has long been home to many self-taught researchers and engineers. If you have an unconventional educational background, we encourage you to apply. Our goal is for this program to be as inclusive and diverse as possible, and we will provide immigration and relocation support to high-potential talent globally.

“We’re going to need the best, most diverse talent and innovative minds out there to achieve our mission,” said Sam Altman, OpenAI’s CEO. “This type of thinker might be at a university, they might be fresh out of high school, working at a cutting-edge tech company or building something on their own. This program is an excellent way for people who are curious, passionate, and skilled to sharpen their focus on AI and machine learning — and to help us invent the future.”

The AI Software Engineering track is a great match for people that have an engineering background and would like to advance to a Software Engineering position in an AI company. We are looking for candidates with engineering experience in fast-paced environments.

“What’s unique about being a software engineer at OpenAI is the scale and novelty of the problems you’re working on every day,” said Christina Kim, a 2021 Scholar who is now working full-time as a Software Engineer on our AppliedAI team. “OpenAI is at the cutting edge of engineering problems. If you’re excited about AI research, you can easily get involved in cross-functional work with a machine learning component by leveraging your engineering skills to further the company’s research.”

The AI Research track is ideal for people with a research background in a scientific non-ML field who would like to transition into a Research Scientist or Research Engineering position. We are looking for a record of achievement in another field such as mathematics, physics or neuroscience.

“This was my foot in the door to get into AI research,” said Christine McLeavey, a 2018 Scholar and Fellow who now manages OpenAI’s Multimodal team. “I had been studying on my own through online courses like deeplearning.ai and fast.ai for a year, and the support from OpenAI gave me the confidence to jump into the field full-time. I learned so much from being around such amazing researchers, and their mentorship had a huge influence on my project (MuseNet).”

“Having a mentor, gaining access to key infrastructure and tooling, and being part of the broader community at OpenAI helped me acquire a research taste, and to get to the interesting research questions a lot faster,” said Jonathan Ward, former Scholar and Fellow who will be returning to OpenAI in 2022 as a full-time Researcher on the Alignment team.

Applications for the Spring cohort in 2022 are open now through January 14, 2022 12AM (PST). Join us for a discussion with a panel of former OpenAI Scholars and Fellows on December 8 to learn more about the program. Add this event to your calendar.

Apply now

OpenAI

MURAL: Multimodal, Multi-task Retrieval Across Languages

Posted by Aashi Jain, AI Resident and Yinfei Yang, Staff Research Scientist, Google Research

For many concepts, there is no direct one-to-one translation from one language to another, and even when there is, such translations often carry different associations and connotations that are easily lost for a non-native speaker. In such cases, however, the meaning may be more obvious when grounded in visual examples. Take, for instance, the word “wedding”. In English, one often associates a bride in a white dress and a groom in a tuxedo, but when translated into Hindi (शादी), a more appropriate association may be a bride wearing vibrant colors and a groom wearing a sherwani. What each person associates with the word may vary considerably, but if they are shown an image of the intended concept, the meaning becomes more clear.

The word “wedding” in English and Hindi conveys different mental images. Images are taken from wikipedia, credited to Psoni2402 (left) and David McCandless (right) with CC BY-SA 4.0 license.

With current advances in neural machine translation and image recognition, it is possible to reduce this sort of ambiguity in translation by presenting a text paired with a supporting image. Prior research has made much progress in learning image–text joint representations for high-resource languages, such as English. These representation models strive to encode the image and text into vectors in a shared embedding space, such that the image and the text describing it are close to each other in that space. For example, ALIGN and CLIP have shown that training a dual-encoder model (i.e., one trained with two separate encoders) on image–text pairs using a contrastive learning loss works remarkably well when provided with ample training data.

Unfortunately, such image–text pair data does not exist at the same scale for the majority of languages. In fact, more than 90% of this type of web data belongs to the top-10 highly-resourced languages, such as English and Chinese, with much less data for under-resourced languages. To overcome this issue, one could either try to manually collect image–text pair data for under-resourced languages, which would be prohibitively difficult due to the scale of the undertaking, or one could seek to leverage pre-existing datasets (e.g., translation pairs) that could inform the necessary learned representations for multiple languages.

In “MURAL: Multimodal, Multitask Retrieval Across Languages”, presented at Findings of EMNLP 2021, we describe a representation model for image–text matching that uses multitask learning applied to image–text pairs in combination with translation pairs covering 100+ languages. This technology could allow users to express words that may not have a direct translation into a target language using images instead. For example, the word “valiha”, refers to a type of tube zither played by the Malagasy people, which lacks a direct translation into most languages, but could be easily described using images. Empirically, MURAL shows consistent improvements over state-of-the-art models, other benchmarks, and competitive baselines across the board. Moreover, MURAL does remarkably well for the majority of the under-resourced languages on which it was tested. Additionally, we discover interesting linguistic correlations learned by MURAL representations.

MURAL Architecture
The MURAL architecture is based on the structure of ALIGN, but employed in a multitask fashion. Whereas ALIGN uses a dual-encoder architecture to draw together representations of images and associated text descriptions, MURAL employs the dual-encoder structure for the same purpose while also extending it across languages by incorporating translation pairs. The dataset of image–text pairs is the same as that used for ALIGN, and the translation pairs are those used for LaBSE.

MURAL solves two contrastive learning tasks: 1) image–text matching and 2) text–text (bitext) matching, with both tasks sharing the text encoder module. The model learns associations between images and text from the image–text data, and learns the representations of hundreds of diverse languages from the translation pairs. The idea is that a shared encoder will transfer the image–text association learned from high-resource languages to under-resourced languages. We find that the best model employs an EfficientNet-B7 image encoder and a BERT-large text encoder, both trained from scratch. The learned representation can be used for downstream visual and vision-language tasks.

The architecture of MURAL depicts dual encoders with a shared text-encoder between the two tasks trained using a contrastive learning loss.

Multilingual Image-to-Text and Text-to-Image Retrieval
To demonstrate MURAL’s capabilities, we choose the task of cross-modal retrieval (i.e., retrieving relevant images given a text and vice versa) and report the scores on various academic image–text datasets covering well-resourced languages, such as MS-COCO (and its Japanese variant, STAIR), Flickr30K (in English) and Multi30K (extended to German, French, Czech), XTD (test-only set with seven well-resourced languages: Italian, Spanish, Russian, Chinese, Polish, Turkish, and Korean). In addition to well-resourced languages, we also evaluate MURAL on the recently published Wikipedia Image–Text (WIT) dataset, which covers 108 languages, with a broad range of both well-resourced (English, French, Chinese, etc.) and under-resourced (Swahili, Hindi, etc.) languages.

MURAL consistently outperforms prior state-of-the-art models, including M3P, UC2, and ALIGN, in both zero-shot and fine-tuned settings evaluated on well-resourced and under-resourced languages. We see remarkable performance gains for under-resourced languages when compared to the state-of-the-art model, ALIGN.

Mean recall on various multilingual image–text retrieval benchmarks. Mean recall is a common metric used to evaluate cross-modal retrieval performance on image–text datasets (higher is better). It measures the Recall@N (i.e., the chance that the ground truth image appears in the first N retrieved images) averaged over six measurements: Image→Text and Text→Image retrieval for N=[1, 5, 10]. Note that XTD scores report Recall@10 for Text→Image retrieval.

Retrieval Analysis
We also analyzed zero-shot retrieved examples on the WIT dataset comparing ALIGN and MURAL for English (en) and Hindi (hi). For under-resourced languages like Hindi, MURAL shows improved retrieval performance compared to ALIGN that reflects a better grasp of the text semantics.

Comparison of the top-5 images retrieved by ALIGN and by MURAL for the Text→Image retrieval task on the WIT dataset for the Hindi text, एक तश्तरी पर बिना मसाले या सब्ज़ी के रखी हुई सादी स्पगॅत्ती”, which translates to the English, “A bowl containing plain noodles without any spices or vegetables”.

Even for Image→Text retrieval in a well-resourced language, like French, MURAL shows better understanding for some words. For example, MURAL returns better results for the query “cadran solaire” (“sundial”, in French) than ALIGN, which doesn’t retrieve any text describing sundials (below).

Comparison of the top-5 text results from ALIGN and from MURAL on the Image→Text retrieval task for the same image of a sundial.

Embeddings Visualization
Previously, researchers have shown that visualizing model embeddings can reveal interesting connections among languages — for instance, representations learned by a neural machine translation (NMT) model have been shown to form clusters based on their membership to a language family. We perform a similar visualization for a subset of languages belonging to the Germanic, Romance, Slavic, Uralic, Finnic, Celtic, and Finno-Ugric language families (widely spoken in Europe and Western Asia). We compare MURAL’s text embeddings with LaBSE’s, which is a text-only encoder.

A plot of LabSE’s embeddings shows distinct clusters of languages influenced by language families. For instance, Romance languages (in purple, below) fall into a different region than Slavic languages (in brown, below). This finding is consistent with prior work that investigates intermediate representations learned by a NMT system.

Visualization of text representations of LaBSE for 35 languages. Languages are color coded based on their genealogical association. Representative languages include: Germanic (red) — German, English, Dutch; Uralic (orange) — Finnish, Estonian; Slavic (brown) — Polish, Russian; Romance (purple) — Italian, Portuguese, Spanish; Gaelic (blue) — Welsh, Irish.

In contrast to LaBSE’s visualization, MURAL’s embeddings, which are learned with a multimodal objective, shows some clusters that are in line with areal linguistics (where elements are shared by languages or dialects in a geographic area) and contact linguistics (where languages or dialects interact and influence each other). Notably, in the MURAL embedding space, Romanian (ro) is closer to the Slavic languages like Bulgarian (bg) and Macedonian (mk), which is in line with the Balkan sprachbund, than it is in LaBSE. Another possible language contact brings Finnic languages, Estonian (et) and Finnish (fi), closer to the Slavic languages cluster. The fact that MURAL pivots on images as well as translations appears to add an additional view on language relatedness as learned in deep representations, beyond the language family clustering observed in a text-only setting.

Visualization of text representations of MURAL for 35 languages. Color coding is the same as the figure above.

Final Remarks
Our findings show that training jointly using translation pairs helps overcome the scarcity of image–text pairs for many under-resourced languages and improves cross-modal performance. Additionally, it is interesting to observe hints of areal linguistics and contact linguistics in the text representations learned by using a multimodal model. This warrants more probing into different connections learned implicitly by multimodal models, such as MURAL. Finally, we hope this work promotes further research in the multimodal, multilingual space where models learn representations of and connections between languages (expressed via images and text), beyond well-resourced languages.

Acknowledgements
This research is in collaboration with Mandy Guo, Krishna Srinivasan, Ting Chen, Sneha Kudugunta, Chao Jia, and Jason Baldridge. We thank Zarana Parekh, Orhan Firat, Yuqing Chen, Apu Shah, Anosh Raj, Daphne Luong, and others who provided feedback for the project. We are also grateful for general support from Google Research teams.

Read More

Sharing the results of sponsored independent research on furthering inclusion across economies

We are pleased to share a series of reports that examine BIPOC entrepreneurship across eight markets. The reports, which were commissioned by Meta, analyze the mechanisms of economic exclusion in these markets, and how digital tools help entrepreneurs establish and grow successful small businesses, helping to level the playing field.

Interviews with eight business owners, as well as a review of existing research and publicly available data, reveal that entrepreneurs from marginalized groups use digital tools to do the following:

  • Find alternative sources of funding and investment, in contrast to more prohibitive traditional channels
  • Reduce start-up, marketing, and other costs when launching and growing their businesses
  • Create new product offerings and innovate for the unmet needs of their co-ethnic consumers

These findings build on previous work by Meta’s Policy Advocacy and Research team and Data for Good team. Preliminary findings from many of our projects, including Digital Tools in Crisis and Recovery, Dynamic Markets, and Global State of Small Business, show that SMB leaders from minority communities in the United States were more likely to turn to digital tools to mitigate the economic effects of the pandemic on their businesses.

We wanted to dig deeper into these results and expand our analysis to other countries, so we commissioned a team of leading experts to assess the state of economic inclusion and the impact of digital tools around the world. Two authors — Dr. Jock Collins, scholar of Australian immigration and indigenous entrepreneurship, and Dr. Mohanbir Sawhney, expert on business innovation, including among underprivileged groups — led the research, alongside a panel of seven regional consulting experts with global and local expertise on minority entrepreneurship.

The resulting white paper and individual country reports shed light on the unique challenges facing BIPOC business leaders, as well how digital technologies help them overcome the social, economic, and cultural barriers that they face in conducting business.

To read the reports, click the link below.

View reports

The post Sharing the results of sponsored independent research on furthering inclusion across economies appeared first on Facebook Research.

Read More

Recognizing the 2021 TensorFlow Contributor Awardees

Posted by the TensorFlow team

TensorFlow wouldn’t be where it is today without its bustling global community of contributors. There are many ways these developers contribute. They write code to improve TensorFlow, teach classes, answer questions on forums, and organize and host events.

We are thankful to every person that’s helped the TensorFlow community over the years. And at this year’s TensorFlow Contributor Summit, we wanted to show thanks by recognizing individuals who went above and beyond on their TensorFlow contributions in 2021.

So without further ado, we are pleased to introduce the TensorFlow Contributor Awardees of 2021!

SIG Leadership Award

Awarded to a highly active SIG

Jason Zaman, SIG Build

Active SIG Award

Awarded to an impactful Special Interest Group (SIG) leader

Sean Morgan, SIG Add-ons

TF Forum Award

Awarded to a helpful TF Forum user with many liked posts and responses

Ekaterina Dranitsyna

Diversity and Inclusion Award

Awarded to the person who made a significant effort to bring diversity into the TensorFlow ecosystem

Merve Noyan

Education Outreach Awards

Awarded to the people who made significant contributions to educational outreach

Gant Laborde

Sandeep Mistry

Community Management Award

Awarded to highly active community leaders

TensorFlow User Group Pune (TFUG Pune)

Yogesh Kulkarni, Shashank Sane, and Aditya Kane

Regional Awards

Awarded to top contributors by geographic region

Margaret Maynard-Reid, Americas

Sayak Paul, South Asia / Oceania

Chansung Park, East Asia

Ruqiya Bin Safi, Middle East / Africa

M. Yusuf Sarıgöz, Europe

Art by Margaret Maynard-Reid
Art by Margaret Maynard-Reid

Thank you again to all the TensorFlow contributors! We look forward to recognizing even more of you next year.

Read More

If I Had a Hammer: Purdue’s Anvil Supercomputer Will See Use All Over the Land

Carol Song is opening a door for researchers to advance science on Anvil, Purdue University’s new AI-ready supercomputer, an opportunity she couldn’t have imagined as a teenager in China.

“I grew up in a tumultuous time when, unless you had unusual circumstances, the only option for high school grads was to work alongside farmers or factory workers, then suddenly I was told I could go to college,” said Song, now the project director of Anvil.

And not just any college. Her scores on a national entrance exam opened the door to Tsinghua University, home to China’s most prestigious engineering school.

Along the way, someone told her computers would be big, so she signed up for computer science before she had ever seen a computer. She learned soon enough.

“We were building hardware from the ground up, designing microinstructions and logic circuits, so I got to understand computers from the inside out,” she said.

Easing Access to Supercomputers

Skip forward a few years to grad school at the University of Illinois when another big door opened.

While working in distributed systems, she was hired as one of the first programmers at the National Center for Supercomputing Applications,  one of the first sites in a U.S. program funding supercomputers that researchers shared.

To make the systems more accessible, she helped develop alternatives to the crude editing tools of the day that displayed one line of a program at a time. And she helped pioneering researchers like Michael Norman create visualizations of their work.

GPUs Add AI to HPC

In 2005, she joined Purdue, where she has helped manage nearly three dozen research projects representing more than $60 million in grants as a senior research scientist in the university’s supercomputing center.

“All that helped when we started defining Anvil. I see researchers’ pain points when they are getting on a new system,” said Song.

Anvil links 1,000 Dell EMC PowerEdge C6525 server nodes with 2,000 of the latest AMD x86 CPUs and 64 NVIDIA A100 Tensor Core GPUs on a NVIDIA Quantum InfiniBand network to handle traditional HPC and new AI workloads.

The system, built by Dell Technologies, will deliver 5.3 petaflops and half a million GPU cycles per year to tens of thousands of researchers across the U.S. working on the National Science Foundation’s XSEDE network.

Anvil Forges Desktop, Cloud Links

To harness that power, Anvil supports interactive user interfaces as well as the batch jobs that are traditional in high performance computing.

“Researchers can use their favorite tools like Jupyter notebooks and remote desktop interfaces so the cluster can look just like in their daily work environment,” she said.

Anvil will also support links to Microsoft Azure, so researchers can access its large datasets and commercial cloud-computing muscle. “It’s an innovative part of this system that will let researchers experiment with creating workflows that span research and commercial environments,” Song said.

Fighting COVID, Exploring AI

More than 30 research teams have already signed up to be early users of Anvil.

One team will apply deep learning to medical images to improve diagnosis of respiratory diseases including COVID-19. Another will build causal and logical check points into neural networks to explore why deep learning delivers excellent results.

“We’ll support a lot of GPU-specific tools like NGC containers for accelerated applications, and as with every new system, users can ask for additional toolkits and libraries they want,” she said.

The Anvil team aims to invite industry collaborations to test new ideas using up to 10 percent of the system’s capacity. “It’s a discretionary use we want to apply strategically to enable projects that wouldn’t happen without such resources,” she said.

Opening Doors for Science and Inclusion

Early users are working on Anvil today and the system will be available for all users in about a month.

Anvil’s opening day has a special significance for Song, one of the few women to act as a lead manager for a national supercomputer site.

Carol Song. project director, Purdue Anvil supercomputer
Carol Song and Purdue’s Anvil supercomputer

“I’ve been fortunate to be in environments where I’ve always been encouraged to do my best and given opportunities,” she said.

“Around the industry and the research computing community there still aren’t a lot of women in leadership roles, so it’s an ongoing effort and there’s a lot of room to do better, but I’m also very enthusiastic about mentoring women to help them get into this field,” she added.

Purdue’s research computing group shares Song’s enthusiasm about getting women into supercomputing. It’s home to one of the first chapters of the international Women in High-Performance Computing organization.

Purdue’s Women in HPC chapter sent an all-female team to a student cluster competition at SC18. It also hosts outside speakers, provides travel support to attend conferences and connects students and early career professionals to experienced mentors like Song.

Pictured at top: Carol Song, Anvil’s principal investigator (PI) and project director along with Anvil co-PIs (from left) Rajesh Kalyanam, Xiao Zhu and Preston Smith. 

The post If I Had a Hammer: Purdue’s Anvil Supercomputer Will See Use All Over the Land appeared first on The Official NVIDIA Blog.

Read More

NVIDIA AI Enterprise Helps Researchers, Hospitals Targeting Cancer Hit the Mark

Whether facilitating cancer screenings, cutting down on false positives, or improving tumor identification and treatment planning, AI is a powerful agent for healthcare innovation and acceleration.

Yet, despite its promise, integrating AI into actual solutions can challenge many IT organizations.

The Netherlands Cancer Institute (NKI), one of the world’s top-rated cancer research and treatment centers, is using the NVIDIA AI Enterprise software suite to test AI workloads on higher-precision 3D cancer scans than are commonly used today.

NKI’s AI model was previously trained on lower-resolution images. But with the higher memory capacity offered by NVIDIA AI Enterprise, its researchers could instead use high-resolution images for training. This improvement helps clinicians better target the size and location of a tumor every time a patient receives treatment.

The NVIDIA AI Enterprise suite that NKI deployed is designed to optimize the development and deployment of AI. It’s certified and supported by NVIDIA to enable hospitals, researchers and IT professionals to run AI workloads on mainstream servers with VMware vSphere in their on-prem data centers and private clouds.

Delivering treatments on virtualized infrastructure means hospitals and research institutions can use the same tools they already work with on existing applications. This helps maximize their investments while making innovations in care more affordable and accessible.

NKI used an AI model to better reconstruct a Cone Beam Computed Tomography (CBCT) thoracic image, resulting in clearer image quality compared to conventional methods.

Speeding Breakthroughs in Healthcare Research 

NKI had gotten off to a quick start with its project on NVIDIA AI Enterprise by using NVIDIA LaunchPad.

The LaunchPad program provides immediate access to optimized software running on accelerated infrastructure to help customers prototype and test data science and AI workloads. This month, the program was extended to nine Equinix locations worldwide.

The NVIDIA AI Enterprise software suite, available in LaunchPad, makes it possible to run advanced AI workloads on mainstream accelerated servers with VMware vSphere, including systems from Dell Technologies, Hewlett Packard Enterprise, Lenovo and many others.

Rhino Health, a federated learning platform powered by NVIDIA FLARE, is available today through NVIDIA AI Enterprise, making it easy for any hospital to leverage Federated learning for AI development and validation. Other organizations, like The American College of Radiology’s AI LAB, are also planning to use the NVIDIA AI Enterprise software.

Researchers at NKI used NVIDIA AI Enterprise, running on the HPE Synergy, a composable software system from Hewlett Packard Enterprise, to build deep learning models by combining the massive 2D and 3D data sources and AI to pinpoint the location of tumors before each radiotherapy treatment session. 

“Doctors could use this solution as an alternative to CT scans on day of treatment to optimize the treatment plan to validate the radiotherapy plan,” said Jonas Teuwen, group leader at the Netherlands Cancer Institute.

Using NVIDIA AI Enterprise, Teuwen’s team in Amsterdam ran their workloads on NVIDIA A100 80GB GPUs in a server hosted in Silicon Valley. Their convolutional neural network was built in less than three months and was trained on less than 300 clinical lung CT scans that were then reconstructed and generalized to head and neck data.

In the future, NKI researchers also hope to translate this work to potential use cases in interventional radiology to repair arteries in cardiac surgeries and dental surgery implants.

Accelerating Hospital AI Deployment With NVIDIA AI Enterprise

NVIDIA AI Enterprise simplifies the AI rollout experience for organizations who host a variety of healthcare and operations applications on virtualized infrastructure. It enables IT administrators to run AI applications like Vyasa and iCAD alongside core hospital applications, streamlining the workflow in an environment they’re already familiar with.

Compute resources can be adjusted with just a few clicks, giving hospitals the ability to transform care for both patients and healthcare providers.

Vyasa, a provider of deep learning analysis tools for healthcare and life sciences, uses NVIDIA AI Enterprise to build applications that can search unstructured content such as patient care records. With the software, Vyasa can develop their deep learning applications faster and help dive through unstructured data and PDFs to assess which patients are at a higher risk. It identifies those who haven’t been in for a check-up in more than a year, and can refine for additional risk factors like age and race.

“NVIDIA AI Enterprise has reduced our deployment times by half thanks to rapid provisioning of platform requirements that eliminate the need to manually download and integrate software packages,” said Frans Lawaetz, CIO at Vyasa. 

Radiologists use iCAD’s innovative ProFound AI software to assist with reading mammograms. These AI solutions help identify cancer earlier, categorize breast density, and accurately assess short-term personalized breast cancer risk based on each woman’s screening mammogram. Running advanced workloads with VMware vSphere is important for iCAD’s healthcare customers as they can easily integrate their data intensive applications into any hospital infrastructure.

A host of other software makers, like the American College of Radiology’s AI LAB and Rhino Health, with its federated learning platform, have begun validating their software on NVIDIA AI Enterprise to ease deployment by integrating into a common healthcare IT infrastructure.

The ability for NVIDIA AI Enterprise to unify the data center for healthcare organizations has sparked the creation of an ecosystem with NVIDIA technology at its heart. The common NVIDIA and VMware infrastructure benefits software vendors and healthcare organizations alike by making the deployment and management of these solutions much easier.

For many healthcare IT and software companies, integrating AI into hospital environments is a top priority. Many NVIDIA Inception partners will be testing the ease of deploying their offerings on NVIDIA AI Enterprise in these types of environments. They include ​​Aidence, Arterys, contextflow, ImageBiopsy Lab, InformAI, MD.ai, methinks.ai, RADLogics, Sciberia, Subtle Medical and VUNO.

NVIDIA Inception is a program that offers go-to-market support, expertise and technology for AI, data science and HPC startups.

Qualified enterprises can apply to experience NVIDIA AI Enterprise in curated, no-cost labs offered on NVIDIA LaunchPad.

Hear more about NVIDIA’s work in healthcare by tuning in to my special address on Nov. 29 at RSNA, the Radiological Society of North America’s annual meeting.

Main image shows how NVIDIA AI Enterprise allows hospital IT administrators to run AI applications alongside core hospital applications, like iCAD Profound AI Software for mammograms.

The post NVIDIA AI Enterprise Helps Researchers, Hospitals Targeting Cancer Hit the Mark appeared first on The Official NVIDIA Blog.

Read More

Federated Learning With FLARE: NVIDIA Brings Collaborative AI to Healthcare and Beyond

NVIDIA is making it easier than ever for researchers to harness federated learning by open-sourcing NVIDIA FLARE, a software development kit that helps distributed parties collaborate to develop more generalizable AI models.

Federated learning is a privacy-preserving technique that’s particularly beneficial in cases where data is sparse, confidential or lacks diversity. But it’s also useful for large datasets, which can be biased by an organization’s data collection methods, or by patient or customer demographics.

NVIDIA FLARE — short for Federated Learning Application Runtime Environment — is the engine underlying NVIDIA Clara Train’s federated learning software, which has been used for AI applications in medical imaging, genetic analysis, oncology and COVID-19 research. The SDK allows researchers and data scientists to adapt their existing machine learning and deep learning workflows to a distributed paradigm.

Making NVIDIA FLARE open source will better empower cutting-edge AI in almost any industry by giving researchers and platform developers more tools to customize their federated learning solutions.

With the SDK, researchers can choose among different federated learning architectures, tailoring their approach for domain-specific applications. And platform developers can use NVIDIA FLARE to provide customers with the distributed infrastructure required to build a multi-party collaboration application.

Flexible Federated Learning Workflows for Multiple Industries 

Federated learning participants work together to train or evaluate AI models without having to pool or exchange each group’s proprietary datasets. NVIDIA FLARE provides different distributed architectures that accomplish this, including peer-to-peer, cyclic and server-client approaches, among others.

Using the server-client technique, where learned model parameters from each participant are sent to a common server and aggregated into a global model, NVIDIA has led federated learning projects that help segment pancreatic tumors, classify breast density in mammograms to inform breast cancer risk, and predict oxygen needs for COVID patients.

The server-client architecture was also used for two federated learning collaborations using NVIDIA FLARE: NVIDIA worked with Roche Digital Pathology researchers to run a successful internal simulation using whole slide images for classification, and with Netherlands-based  Erasmus Medical Center for an AI application that identifies genetic variants associated with schizophrenia cases.

But not every federated learning application is suited to the server-client approach. By supporting additional architectures, NVIDIA FLARE will make federated learning accessible to a wider range of applications. Potential use cases include helping energy companies analyze seismic and wellbore data, manufacturers optimize factory operations and financial firms improve fraud detection models.

NVIDIA FLARE Integrates With Healthcare AI Platforms

NVIDIA FLARE can integrate with existing AI initiatives, including the open-source MONAI framework for medical imaging.

“Open-sourcing NVIDIA FLARE to accelerate federated learning research is especially important in the healthcare sector, where access to multi-institutional datasets is crucial, yet concerns around patient privacy can limit the ability to share data,” said Dr. Jayashree Kalapathy, associate professor of radiology at Harvard Medical School and leader of the MONAI community’s federated learning working group. “We are excited to contribute to NVIDIA FLARE and continue the integration with MONAI to push the frontiers of medical imaging research.”

NVIDIA FLARE will also be used to power federated learning solutions at: 

  • American College of Radiology (ACR): The medical society has worked with NVIDIA on federated learning studies that apply AI to radiology images for breast cancer and COVID-19 applications. It plans to distribute NVIDIA FLARE in the ACR AI-LAB, a software platform that is available to the society’s tens of thousands of members.
  • Flywheel: The company’s Flywheel Exchange platform enables users to access and share data and algorithms for biomedical research, manage federated projects for analysis and training, and choose their preferred federated learning solution — including NVIDIA FLARE.
  • Taiwan Web Service Corporation: The company offers a GPU-powered MLOps platform that enables customers to run federated learning based on NVIDIA FLARE. Five medical imaging projects are currently being conducted on the company’s private cluster, each with several participating hospitals.
  • Rhino Health: The partner and member of the NVIDIA Inception program has integrated NVIDIA FLARE into its federated learning solution, which is helping researchers at Massachusetts General Hospital develop an AI model that more accurately diagnoses brain aneurysms, and experts at the National Cancer Institute’s Early Detection Research Network develop and validate medical imaging AI models that identify early signs of pancreatic cancer.

“To collaborate effectively and efficiently, healthcare researchers need a common platform for AI development without the risk of breaching patient privacy,” said Dr. Ittai Dayan, founder of Rhino Health. “Rhino Health’s ‘Federated Learning as a Platform’ solution, built with NVIDIA FLARE, will be a useful tool to help accelerate the impact of healthcare AI.”

Get started with federated learning by downloading NVIDIA FLARE. Hear more about NVIDIA’s work in healthcare by tuning in to a special address on Nov. 29 at 6 p.m. CT by David Niewolny, director of healthcare business development at NVIDIA, at RSNA, the Radiological Society of North America’s annual meeting.

Subscribe to NVIDIA healthcare news here

The post Federated Learning With FLARE: NVIDIA Brings Collaborative AI to Healthcare and Beyond appeared first on The Official NVIDIA Blog.

Read More

Improve the return on your marketing investments with intelligent user segmentation in Amazon Personalize

Today, we’re excited to announce intelligent user segmentation powered by machine learning (ML) in Amazon Personalize, a new way to deliver personalized experiences to your users and run more effective campaigns through your marketing channels.

Traditionally, user segmentation depends on demographic or psychographic information to sort users into predefined audiences. More advanced techniques look to identify common behavioral patterns in the customer journey (such as frequent site visits, recent purchases, or cart abandonment) using business rules to derive users’ intent. These techniques rely on assumptions about the users’ preferences and intentions that limit their scalability, don’t automatically learn from changing user behaviors, and don’t offer user experiences personalized for each user. User segmentation in Amazon Personalize uses ML techniques, developed and perfected at Amazon, to learn what is relevant to users. Amazon Personalize automatically identifies high propensity users without the need to develop and maintain an extensive and brittle catalog of rules. This means you can create more effective user segments that scale with your catalog and learn from your users’ changing behavior to deliver what matters to them.

Amazon Personalize enables developers to build personalized user experiences with the same ML technology used by Amazon with no ML expertise required. We make it easy for developers to build applications capable of delivering a wide array of personalization experiences. You can start creating user segments quickly with the Amazon Personalize API or AWS Management Console and only pay for what you use, with no minimum fees or upfront commitments. All data is encrypted to be private and secure, and is only used to create your user segments.

This post walks you through how to use Amazon Personalize to segment your users based on preferences for grocery products using an Amazon Prime Pantry dataset.

Overview of solution

We’re introducing two new recipes that segment your users based on their interest in different product categories, brands, and more. Our item affinity recipe (aws-item-affinity) identifies users based on their interest in the individual items in your catalog, such as a movie, song, or product. The item attribute affinity recipe (aws-item-attribute) identifies users based on the attributes of items in your catalog, such as genre or brand. This allows you to better engage users with your marketing campaigns and improve retention through targeted messaging.

The notebook that accompanies this post demonstrates how to use the aws-item-affinity and aws-item-attribute recipe to create user segments based on their preferences for grocery products in an Amazon Prime Pantry dataset. We use one dataset group that contains user-item interaction data and item metadata. We use these datasets to train solutions using the two recipes and create user segments in batch.

To test the performance of the solution, we split the interactions data into a training set and test set. The Amazon Prime Pantry dataset has approximately 18 years of interaction data from August 9, 2000, to October 5, 2018, with approximately 1.7 million interactions. We hold out 5% of the most recent interactions and train on the remaining 95%. This results in a split where we use interactions from August 9, 2000, through February 1, 2018, to train the solution and use the remaining 8 months of interactions to simulate future activity as ground truth.

Results

When reproducing these tests in the notebook, your results may vary slightly. This is because when training, the solution the parameters of the underlying models are randomly initialized.

Let’s first review the results by looking at a few examples. We ran queries on three items, and identified 10 users that have a high propensity to engage with the items. We then look at the users’ shopping histories to assess if they would likely be interested in the queried product.

The following table shows the results of a segmentation query on gingerbread coffee, an item we might want to promote for the holiday season. Each row in the table shows the last three purchases of the 10 users returned from the query. Most of the users we identified are clearly coffee drinkers, having recently purchased coffee and coffee creamers. Interestingly, the item we queried on is a whole bean coffee, not a ground coffee. We see in the item histories that, where the information is available, the users have recently purchased whole bean coffee.

Gingerbread Coffee, 1 lb Whole Bean FlavorSeal Vacuum Bag: Bite into a freshly baked Gingerbread Coffee
USER_ID Last Three Purchases
A1H3ATRIQ098I7 Brew La La Red Velvet Cupcake Coffee Ola’s Exotic Super Premium Coffee Organic Uganda B Coffee Masters Gourmet Coffee
ANEDXRFDZDL18 Pepperidge Farm Goldfish Crackers Boston Baked Beans (1) 5.3 Oz Theater Box Sizecont Boost Simply Complete Nutritional Drink
APHFL4MDJRGWB Dunkin’ Donuts Original Blend Ground Coffee Coffee-Mate Coffee Mix Folgers Gourmet Selections Coconut Cream Pie Flavo
ANX42D33MNOVP The Coffee Fool Fool’s House American Don Francisco’s Hawaiian Hazelnut Don Francisco’s French Roast Coffee
A2NLJJVA0IEK2S Coffee Masters Flavored Coffee Lays 15pk Hickory Sticks Original (47g / 1.6oz per Albanese Confectionery Sugar Free Gummy Bears
A1GDEQIGFPRBNO Christopher Bean Coffee Flavored Ground Coffee Cameron’s French Vanilla Almond Whole Bean Coffee Cameron’s Coffee Roasted Whole Bean Coffee
A1MDO8RZCZ40B0 Master Chef Ground Coffee New England Ground Coffee Maxwell House Wake Up Roast Medium Coffee
A2LK2DENORQI8S The Bean Coffee Company Organic Holiday Bean (Vani Lola Savannah Angel Dust Ground New England Coffee Blueberry Cobbler
AGW1F5N8HV3AS New England Coffee Colombian Kirkland Signature chicken breast Lola Savannah Banana Nut Whole Bean
A13YHYM6FA6VJO Lola Savannah Triple Vanilla Whole Bean Lola Savannah Vanilla Cinnamon Pecan Whole Bean Pecan Maple Nut

The next table shows a segmentation query on hickory liquid smoke, a seasoning used for barbecuing and curing bacon. We see a number of different grocery products that might accompany barbecue in the users’ recent purchases: barbecue sauces, seasonings, and hot sauce. Two of the users recently purchased Prague Powder No. 1 Pink Curing Salt, a product also used for curing bacon. We may have missed these two users if we had relied on rules to identify people interested in grilling.

Wright’s Natural Hickory Seasoning Liquid Smoke, 128 Ounce This seasoning is produced by burning fresh cut hickory chips, then condensing the smoke into a liquid form.
USER_ID Last Three Purchases
A1MHK19QSCV8SY Hoosier Hill Farm Prague Powder No.1 Pink Curing S APPLE CIDER VINEGAR Fleischmann’s Instant Dry Yeast 1lb bagDry Yeast.M
A3G5P0SU1AW2DO Wright’s Natural Hickory Seasoning Liquid Smoke Eight O’Clock Whole Bean Coffee Kitchen Bouquet Browning and Seasoning Sauce
A2WW9T8EEI8NU4 Hidden Valley Dips Mix Creamy Dill .9 oz Packets ( Frontier Garlic Powder Wolf Chili Without Beans
A2TEJ1S0SK7ZT Black Tai Salt Co’s – (Food Grade) Himalayan Cryst Marukan Genuine Brewed Rice Vinegar Unseasoned Cheddar Cheese Powder
A3MPY3AGRMPCZL Wright’s Natural Hickory Seasoning Liquid Smoke San Francisco Bay OneCup Fog Chaser (120 Count) Si Kikkoman Soy Sauce
A2U77Z3Z7DC9T9 Food to Live Yellow Mustard Seeds (Kosher) 5 Pound 100 Sheets (6.7oz) Dried Kelp Seaweed Nori Raw Uns SB Oriental Hot Mustard Powder
A2IPDJISO5T6AX Angel Brand Oyster Sauce Bullhead Barbecue Sauce ONE ORGANIC Sushi Nori Premium Roasted Organic Sea
A3NDGGX7CWV8RT Frontier Mustard Seed Da Bomb Ghost Pepper HOT SaucesWe infused our hot Starwest Botanicals Organic Rosemary Leaf Whole
A3F7NO1Q3RQ9Y0 Yankee Traders Brand Whole Allspice Aji No Moto Ajinomoto Monosodium Glutamate Umami S Hoosier Hill Farm Prague Powder No.1 Pink Curing S
A3JKI7AWYSTILO Lalah’s Heated Indian Curry Powder 3 Lb LargeCurry Ducal Beans Black Beans with Cheese Emerald Nuts Whole Cashews

Our third example shows a segmentation query on a decoration used to top cakes. We see that the users identified are not only bakers, but are also clearly interested in decorating their baked goods. We see recent purchases like other cake toppers, edible decorations, and fondant (an icing used to sculpt cakes).

Letter C – Swarovski Crystal Monogram Wedding Cake Topper Letter, Jazz up your cakes with a sparkling monogram from our Sparkling collection! These single letter monograms are silver plated covered in crystal rhinestones and come in several sizes for your convenience.
USER_ID Last Three Purchases
A3RLEN577P4E3M The Republic Of Tea Alyssa’s Gluten Free Oatmeal Cookies – Pack of 4. Double Honey Filled Candies
AOZ0D3AGVROT5 Sea Green Disco Glitter Dust Christmas Green Disco Glitter Dust Baby Green Disco Glitter Dust
AC7O52PQ4HPYR Rhinestone Cake Topper Number 7 by otherThis delic Rhinestone Cake Topper Number 5This delicate and h Rhinestone Cake Topper Number 8 by otherThis delic
ALXKY9T83C4Z6 Heart Language of Love Bride and Groom White Weddi Bliss Cake Topper by Lenox (836473)It’s a gift tha First Dance Bride and Groom Wedding Cake TopperRom
A2XERDJ6I2K38U Egyptian Gold Luster Dust Kellogg’s Rice Krispies Treats Wilton Decorator Preferred Green Fondant
A1474SH2RB49MP Assorted Snowflake Sugar Decorations Disney Movie Darice VL3L Mirror Acrylic Initial Letter Cake Top Edible Snowflakes Sugar Decorations (15 pc).
A24E9YGY3V94N8 TOOGOO(R) Double-Heart Cake Topper Decoration for Custom Personalized Mr Mrs Wedding Cake Topper Wit Jacobs Twiglets 6 Pack Jacobs Twiglets are one of
A385P0YAW6U5J3 Tinksky Wedding Cake Topper God Gave Me You Sparkl Sweet Sixteen Cake Topper 16th Birthday Cake Toppe Catching the Big One DecoSet Cake DecorationReel i
A3QW120I2BY1MU Golda’s Kitchen Acetate Cake Collars – 4. Twinings of London English Breakfast Tea K-Cups fo Chefmaster by US Cake Supply 9-Ounce Airbrush Clea
A3DCP979LU7CTE DecoPac Heading for The Green DecoSet Cake TopperL Rhinestne Cake Topper Number 90This delicate and h Rhinestone Cake Topper Letter KThis delicate and h

These three examples make sense based on our editorial judgement, but to truly assess the performance of the recipe, we need to analyze more of the results. To do this broader assessment, we run the aws-item-affinity solution on 500 randomly selected items that appear in the test set to query a list of 2,262 users (1% of the users in the dataset). We then use the test set to assess how frequently the 2,262 users purchased the items during the test period. For comparison, we also assess how frequently 2,262 of the most active users purchased the items during the test period. The following table shows that the aws-item-affinity solution is four times better at identifying users that would purchase a given item.

Test Metrics
Hits Recall
Personalize – Item Affinity 0.2880 0.1297
Active User Baseline 0.0720 0.0320

Although these results are informative, they’re not a perfect reflection of the performance of the recipe because the user segmentation wasn’t used to promote the items which users later interacted with. The best way to measure performance is an online A/B test—running a marketing campaign on a list of users derived from the aws-item-affinity solution alongside a set of the most active users to measure the difference in engagement.

Conclusion

Amazon Personalize now makes it easy to run more intelligent user segmentation at scale, without having to maintain complex sets of rules or relying on broad assumptions about the preferences of your users. This allows you to better engage users with your marketing campaigns and improve retention through targeted messaging.

To learn more about Amazon Personalize, visit the product page.


About the Authors

Daniel Foley is a Senior Product Manager for Amazon Personalize. He is focused on building applications that leverage artificial intelligence to solve our customers’ largest challenges. Outside of work, Dan is an avid skier and hiker.

Debarshi Raha is a Senior Software Engineer for Amazon Personalize. He is passionate about building AI-based personalization systems at scale. In his spare time, he enjoys traveling and photography.

Ge Liu is an Applied Scientist at AWS AI Labs working on developing next generation recommender system for Amazon Personalize. Her research interests include Recommender System, Deep Learning, and Reinforcement Learning.

Haizhou Fu is a senior software engineer on the Amazon Personalize team working on designing and building recommendation systems and solutions for different industries. Outside of his work, he loves playing soccer, basketball and watching movies, reading and learning about physics, especially theories related to time and space.

Read More