A gentle introduction to automated reasoning

Meet Amazon Science’s newest research area.

This week, Amazon Science added automated reasoning to its list of research areas. We made this change because of the impact that automated reasoning is having here at Amazon. For example, Amazon Web Services’ customers now have direct access to automated-reasoning-based features such as IAM Access Analyzer, S3 Block Public Access, or VPC Reachability Analyzer. We also see Amazon development teams integrating automated-reasoning tools into their development processes, raising the bar on the security, durability, availability, and quality of our products.

The goal of this article is to provide a gentle introduction to automated reasoning for the industry professional who knows nothing about the area but is curious to learn more. All you will need to make sense of this article is to be able to read a few small C and Python code fragments. I will refer to a few specialist concepts along the way, but only with the goal of introducing them in an informal manner. I close with links to some of our favorite publicly available tools, videos, books, and articles for those looking to go more in-depth.

Let’s start with a simple example. Consider the following C function:

bool f(unsigned int x, unsigned int y) {
   return (x+y == y+x);
}

Take a few moments to answer the question “Could f ever return false?” This is not a trick question: I’ve purposefully used a simple example to make a point.

To check the answer with exhaustive testing, we could try executing the following doubly nested test loop, which calls f on all possible pairs of values of the type unsigned int:

#include<stdio.h>
#include<stdbool.h>
#include<limits.h>

bool f(unsigned int x, unsigned int y) {
   return (x+y == y+x);
}

void main() {
   for (unsigned int x=0;1;x++) {
      for (unsigned int y=0;1;y++) {
         if (!f(x,y)) printf("Error!\n");
         if (y==UINT_MAX) break;
      }
      if (x==UINT_MAX) break;
   }
}

Unfortunately, even on modern hardware, this doubly nested loop will run for a very long time. I compiled it and ran it on a 2.6 GHz Intel processor for over 48 hours before giving up.

Why does testing take so long? Because UINT_MAX is typically 4,294,967,295, there are 18,446,744,065,119,617,025 separate f calls to consider. On my 2.6 GHz machine, the compiled test loop called f approximately 430 million times a second. But to test all 18 quintillion cases at this performance, we would need over 1,360 years.

When we show the above code to industry professionals, they almost immediately work out that f can't return false as long as the underlying compiler/interpreter and hardware are correct. How do they do that? They reason about it. They remember from their school days that x + y can be rewritten as y + x and conclude that f always returns true.

Re:Invent 2021 keynote address by Peter DeSantis, senior vice president for utility computing at Amazon Web Services
Skip to 15:49 for a discussion of Amazon Web Services' work on automated reasoning.

An automated reasoning tool does this work for us: it attempts to answer questions about a program (or a logic formula) by using known techniques from mathematics. In this case, the tool would use algebra to deduce that x + y == y + x can be replaced with the simple expression true.

Automated-reasoning tools can be incredibly fast, even when the domains are infinite (e.g., unbounded mathematical integers rather than finite C ints). Unfortunately, the tools may answer “Don’t know” in some instances. We'll see a famous example of that below.

The science of automated reasoning is essentially focused on driving the frequency of these “Don’t know” answers down as far as possible: the less often the tools report "Don't know" (or time out while trying), the more useful they are.

Today’s tools are able to give answers for programs and queries where yesterday’s tools could not. Tomorrow’s tools will be even more powerful. We are seeing rapid progress in this field, which is why at Amazon, we are increasingly getting so much value from it. In fact, we see automated reasoning forming its own Amazon-style virtuous cycle, where more input problems to our tools drive improvements to the tools, which encourages more use of the tools.

A slightly more complex example. Now that we know the rough outlines of what automated reasoning is, the next small example gives a slightly more realistic taste of the sort of complexity that the tools are managing for us.

void g(int x, int y) {
   if (y > 0)
      while (x > y)
         x = x - y;
}

Or, alternatively, consider a similar Python program over unbounded integers:

def g(x, y):
   assert isinstance(x, int) and isinstance(y, int)
   if y > 0:
      while x > y:
         x = x - y

Try to answer this question: “Does g always eventually return control back to its caller?”

When we show this program to industry professionals, they usually figure out the right answer quickly. A few, especially those who are aware of results in theoretical computer science, sometimes mistakenly think that we can't answer this question, with the rationale “This is an example of the halting problem, which has been proved insoluble”. In fact, we can reason about the halting behavior for specific programs, including this one. We’ll talk more about that later.

Here’s the reasoning that most industry professionals use when looking at this problem:

  1. In the case where y is not positive, execution jumps to the end of the function g. That’s the easy case.
  2. If, in every iteration of the loop, the value of the variable x decreases, then eventually, the loop condition x > y will fail, and the end of g will be reached.
  3. The value of x always decreases only if y is always positive, because only then does the update to x (i.e., x = x - y) decrease x. But y’s positivity is established by the conditional expression, so x always decreases.

The experienced programmer will usually worry about underflow in the x = x - y command of the C program but will then notice that x > y before the update to x and thus cannot underflow.

If you carried out the three steps above yourself, you now have a very intuitive view of the type of thinking an automated-reasoning tool is performing on our behalf when reasoning about a computer program. There are many nitty-gritty details that the tools have to face (e.g., heaps, stacks, strings, pointer arithmetic, recursion, concurrency, callbacks, etc.), but there’s also decades of research papers on techniques for handling these and other topics, along with various practical tools that put these ideas to work.

Policy-code.gif
Automated reasoning can be applied to both policies (top) and code (bottom). In both cases, an essential step is reasoning about what's always true.

The main takeaway is that automated-reasoning tools are usually working through the three steps above on our behalf: Item 1 is reasoning about the program’s control structure. Item 2 is reasoning about what is eventually true within the program. Item 3 is reasoning about what is always true in the program.

Note that configuration artifacts such as AWS resource policies, VPC network descriptions, or even makefiles can be thought of as code. This viewpoint allows us to use the same techniques we use to reason about C or Python code to answer questions about the interpretation of configurations. It’s this insight that gives us tools like IAM Access Analyzer or VPC Reachability Analyzer.

An end to testing?

As we saw above when looking at f and g, automated reasoning can be dramatically faster than exhaustive testing. With tools available today, we can show properties of f or g in milliseconds, rather than waiting lifetimes with exhaustive testing.

Can we throw away our testing tools now and just move to automated reasoning? Not quite. Yes, we can dramatically reduce our dependency on testing, but we will not be completely eliminating it any time soon, if ever. Consider our first example:

bool f(unsigned int x, unsigned int y) {
   return (x + y == y + x);
}

Recall the worry that a buggy compiler or microprocessor could in fact cause an executable program constructed from this source code to return false. We might also need to worry about the language runtime. For example, the C math library or the Python garbage collector might have bugs that cause a program to misbehave.

What’s interesting about testing, and something we often forget, is that it’s doing much more than just telling us about the C or Python source code. It’s also testing the compiler, the runtime, the interpreter, the microprocessor, etc. A test failure could be rooted in any of those tools in the stack.

Automated reasoning, in contrast, is usually applied to just one layer of that stack — the source code itself, or sometimes the compiler or the microprocessor. What we find so valuable about reasoning is it allows us to clearly define both what we do know and what we do not know about the layer under inspection.

Furthermore, the models of the surrounding environment (e.g., the compiler or the procedure calling our procedure) used by the automated-reasoning tool make our assumptions very precise. Separating the layers of the computational stack helps make better use of our time, energy, and money and the capabilities of the tools today and tomorrow.

Unfortunately, we will almost always need to make assumptions about something when using automated reasoning — for example, the principles of physics that govern our silicon chips. Thus, testing will never be fully replaced. We will want to perform end-to-end testing to try and validate our assumptions as best we can.

An impossible program

I previously mentioned that automated-reasoning tools sometimes return “Don’t know” rather than “yes” or “no”. They also sometimes run forever (or time out), thus never returning an answer. Let’s look at the famous "halting problem" program, in which we know tools cannot return “yes” or “no”.

Imagine that we have an automated-reasoning API, called terminates, that returns “yes” if a C function always terminates or “no” when the function could execute forever. As an example, we could build such an API using the tool described here (shameless self-promotion of author’s previous work). To get the idea of what a termination tool can do for us, consider two basic C functions, g (from above),

void g(int x, int y) {
   if (y > 0)
      while (x > y)
         x = x - y;
}

and g2:

void g2(int x, int y) {
   while (x > y)
      x = x - y;
}

For the reasons we have already discussed, the function g always returns control back to its caller, so terminates(g) should return true. Meanwhile, terminates(g2) should return false because, for example, g2(5, 0) will never terminate.

Now comes the difficult function. Consider h:

void h() {
   if terminates(h) while(1){}
}

Notice that it's recursive. What’s the right answer for terminates(h)? The answer cannot be "yes". It also cannot be "no". Why?

Imagine that terminates(h) were to return "yes". If you read the code of h, you’ll see that in this case, the function does not terminate because of the conditional statement in the code of h that will execute the infinite loop while(1){}. Thus, in this case, the terminates(h) answer would be wrong, because h is defined recursively, calling terminates on itself.

Similarly, if terminates(h) were to return "no", then h would in fact terminate and return control to its caller, because the if case of the conditional statement is not met, and there is no else branch. Again, the answer would be wrong. This is why the “Don’t know” answer is actually unavoidable in this case.

The program h is a variation of examples given in Turing’s famous 1936 paper on decidability and Gödel’s incompleteness theorems from 1931. These papers tell us that problems like the halting problem cannot be “solved”, if bysolved” we mean that the solution procedure itself always terminates and answers either “yes” or “no” but never “Don’t know”. But that is not the definition of “solved” that many of us have in mind. For many of us, a tool that sometimes times out or occasionally returns “Don’t know” but, when it gives an answer, always gives the right answer is good enough.

This problem is analogous to airline travel: we know it’s not 100% safe, because crashes have happened in the past, and we are sure that they will happen in the future. But when you land safely, you know it worked that time. The goal of the airline industry is to reduce failure as much as possible, even though it’s in principle unavoidable.

To put that in the context of automated reasoning: for some programs, like h, we can never improve the tool enough to replace the "Don't know" answer. But there are many other cases where today's tools answer "Don't know", but future tools may be able to answer "yes" or "no". The modern scientific challenge for automated-reasoning subject-matter experts is to get the practical tools to return “yes” or “no” as often as possible. As an example of current work, check out CMU professor and Amazon scholar Marijn Heule and his quest to solve the Collatz termination problem.

Another thing to keep in mind is that automated-reasoning tools are regularly trying to solve “intractable” problems, e.g., problems in the NP complexity class. Here, the same thinking applies that we saw in the case of the halting problem: automated-reasoning tools have powerful heuristics that often work around the intractability problem for specific cases, but those heuristics can (and sometimes do) fail, resulting in “Don’t know” answers or impractically long execution time. The science is to improve the heuristics to minimize that problem.

Nomenclature

A host of names are used in the scientific literature to describe interrelated topics, of which automated reasoning is just one. Here’s a quick glossary:

  • logic is a formal and mechanical system for defining what is true and untrue. Examples: propositional logic or first-order logic.
  • theorem is a true statement in logic. Example: the four-color theorem.
  • proof is a valid argument in logic of a theorem. Example: Gonthier's proof of the four-color theorem
  • mechanical theorem prover is a semi-automated-reasoning tool that checks a machine-readable expression of a proof often written down by a human. These tools often require human guidance. Example: HOL-light, from Amazon researcher John Harrison
  • Formal verification is the use of theorem proving when applied to models of computer systems to prove desired properties of the systems. Example: the CompCert verified C compiler
  • Formal methods is the broadest term, meaning simply the use of logic to reason formally about models of systems. 
  • Automated reasoning focuses on the automation of formal methods. 
  • semi-automated-reasoning tool is one that requires hints from the user but still finds valid proofs in logic. 

As you can see, we have a choice of monikers when working in this space. At Amazon, we’ve chosen to use automated reasoning, as we think it best captures our ambition for automation and scale. In practice, some of our internal teams use both automated and semi-automated reasoning tools, because the scientists we've hired can often get semi-automated reasoning tools to succeed where the heuristics in fully automated reasoning might fail. For our externally facing customer features, we currently use only fully automated approaches.

Next steps

In this essay, I’ve introduced the idea of automated reasoning, with the smallest of toy programs. I haven’t described how to handle realistic programs, with heap or concurrency. In fact, there are a wide variety of automated-reasoning tools and techniques, solving problems in all kinds of different domains, some of them quite narrow. To describe them all and the many branches and sub-disciplines of the field (e.g. “SMT solving”, “higher-order logic theorem proving”, “separation logic”) would take thousands of blogs posts and books.

Automated reasoning goes back to the early inventors of computers. And logic itself (which automated reasoning attempts to solve) is thousands of years old. In order to keep this post brief, I’ll stop here and suggest further reading. Note that it’s very easy to get lost in the weeds reading depth-first into this area, and you could emerge more confused than when you started. I encourage you to use a bounded depth-first search approach, looking sequentially at a wide variety of tools and techniques in only some detail and then moving on, rather than learning only one aspect deeply.

Suggested books:

International conferences/workshops:

Tool competitions:

Some tools:

Interviews of Amazon staff about their use of automated reasoning:

AWS Lectures aimed at customers and industry:

AWS talks aimed at the automated-reasoning science community:

AWS blog posts and informational videos:

Some course notes by Amazon Scholars who are also university professors:

A fun deep track:

Some algorithms found in the automated theorem provers we use today date as far back as 1959, when Hao Wang used automated reasoning to prove the theorems from Principia Mathematica.

Research areas

Related content

LU, Luxembourg
Have you ever ordered a product on Amazon and when that box with the smile arrived you wondered how it got to you so fast? Have you wondered where it came from and how much it cost Amazon to deliver it to you? We are looking for a Senior Data Scientist who will be responsible to develop cutting-edge scientific solutions to optimize our Pan-European fulfillment strategy, to maximize our Customer Experience and minimize our cost and carbon footprint. You will partner with the worldwide scientific community to help design the optimal fulfillment strategy for Amazon. You will also collaborate with technical teams to develop optimization tools for network flow planning and execution systems. Finally, you will also work with business and operational stakeholders to influence their strategy and gather inputs to solve problems. To be successful in the role, you will need deep analytical skills and a strong scientific background. The role also requires excellent communication skills, and an ability to influence across business functions at different levels. You will work in a fast-paced environment that requires you to be detail-oriented and comfortable in working with technical, business and technical teams. Key job responsibilities - Design and develop mathematical models to optimize inventory placement and product flows. - Design and develop statistical and optimization models for planning Supply Chain under uncertainty. - Manage several, high impact projects simultaneously. - Consult and collaborate with business and technical stakeholders across multiple teams to define new opportunities to optimize our Supply Chain. - Communicate data-driven insights and recommendations to diverse senior stakeholders through technical and/or business papers.
IN, HR, Gurugram
Our customers have immense faith in our ability to deliver packages timely and as expected. A well planned network seamlessly scales to handle millions of package movements a day. It has monitoring mechanisms that detect failures before they even happen (such as predicting network congestion, operations breakdown), and perform proactive corrective actions. When failures do happen, it has inbuilt redundancies to mitigate impact (such as determine other routes or service providers that can handle the extra load), and avoids relying on single points of failure (service provider, node, or arc). Finally, it is cost optimal, so that customers can be passed the benefit from an efficiently set up network. Amazon Shipping is hiring Applied Scientists to help improve our ability to plan and execute package movements. As an Applied Scientist in Amazon Shipping, you will work on multiple challenging machine learning problems spread across a wide spectrum of business problems. You will build ML models to help our transportation cost auditing platforms effectively audit off-manifest (discrepancies between planned and actual shipping cost). You will build models to improve the quality of financial and planning data by accurately predicting ship cost at a package level. Your models will help forecast the packages required to be pick from shipper warehouses to reduce First Mile shipping cost. Using signals from within the transportation network (such as network load, and velocity of movements derived from package scan events) and outside (such as weather signals), you will build models that predict delivery delay for every package. These models will help improve buyer experience by triggering early corrective actions, and generating proactive customer notifications. Your role will require you to demonstrate Think Big and Invent and Simplify, by refining and translating Transportation domain-related business problems into one or more Machine Learning problems. You will use techniques from a wide array of machine learning paradigms, such as supervised, unsupervised, semi-supervised and reinforcement learning. Your model choices will include, but not be limited to, linear/logistic models, tree based models, deep learning models, ensemble models, and Q-learning models. You will use techniques such as LIME and SHAP to make your models interpretable for your customers. You will employ a family of reusable modelling solutions to ensure that your ML solution scales across multiple regions (such as North America, Europe, Asia) and package movement types (such as small parcel movements and truck movements). You will partner with Applied Scientists and Research Scientists from other teams in US and India working on related business domains. Your models are expected to be of production quality, and will be directly used in production services. You will work as part of a diverse data science and engineering team comprising of other Applied Scientists, Software Development Engineers and Business Intelligence Engineers. You will participate in the Amazon ML community by authoring scientific papers and submitting them to Machine Learning conferences. You will mentor Applied Scientists and Software Development Engineers having a strong interest in ML. You will also be called upon to provide ML consultation outside your team for other problem statements. If you are excited by this charter, come join us!
US, WA, Bellevue
Amazon SCOT OIH (Supply Chain Optimization Technology - Optimal Inventory Health) team owns inventory health management for Retail worldwide. We use a dynamic programming model to maximize the net present value of inventory driving actions such as pricing markdowns, deals, removals, coupons etc. Our team, the OIH Insights Team energize and empower OIH business with the clarity and conviction required to make impactful business decisions through the generation of actionable and explainable insights, we do so through the following mechanisms: -- Transforming raw, complex datasets into intuitive, and actionable insights that impact OIH strategy and accelerate business decision making. -- Building and maintaining modular, scalable data models that provide the generality, flexibility, intuitiveness, and responsiveness required for seamless self-service insights. -- Generating deeper insights that drive competitive advantage using statistical modeling and machine learning. As a data scientist in the team, you can contribute to each layers of a data solution – you work closely with business intelligence engineers and product managers to obtain relevant datasets and prototype predictive analytic models, you team up with data engineers and software development engineers to implement data pipeline to productionize your models, and review key results with business leaders and stakeholders. Your work exhibits a balance between scientific validity and business practicality. You will be diving deep in our data and have a strong bias for action to quickly produce high quality data analyses with clear findings and recommendations. The ideal candidate is self-motivated, has experience in applying technical knowledge to a business context, can turn ambiguous business questions into clearly defined problems, can effectively collaborate with research scientists, software development engineers, and product managers, and deliver results that meet high standards of data quality, security, and privacy. Key job responsibilities 1. Define and conduct experiments to optimize Long Term Free Cash Flow for Amazon Retail inventory, and communicate insights and recommendations to product, engineering, and business teams 2. Interview stakeholders to gather business requirements and translate them into concrete requirement for data science projects 3. Build models that forecast growth and incorporate inputs from product, engineering, finance and marketing partners 4. Apply data science techniques to automatically identify trends, patterns, and frictions of product life cycle, seasonality, etc 5. Work with data engineers and software development engineers to deploy models and experiments to production 6. Identify and recommend opportunities to automate systems, tools, and processes
US, WA, Seattle
Amazon Advertising is one of Amazon's fastest growing and most profitable businesses. Amazon's advertising portfolio helps merchants, retail vendors, and brand owners succeed via native advertising, which grows incremental sales of their products sold through Amazon. The primary goals are to help shoppers discover new products they love, be the most efficient way for advertisers to meet their business objectives, and build a sustainable business that continuously innovates on behalf of customers. Our products and solutions are strategically important to enable our Retail and Marketplace businesses to drive long-term growth. We deliver billions of ad impressions and millions of clicks and break fresh ground in product and technical innovations every day! As an Applied Scientist on this team, you will: - Be a strong contributor to Machine Learning; lending effort within this team and across other teams. - Perform hands-on analysis and modeling of enormous data sets to develop insights that increase traffic monetization and merchandise sales, without compromising the shopper experience. - Drive end-to-end Machine Learning projects that have a high degree of ambiguity, scale, complexity. - Build machine learning models, perform proof-of-concept, experiment, optimize, and deploy your models into production; work closely with software engineers to assist in productionizing your ML models. - Run A/B experiments, gather data, and perform statistical analysis. - Establish scalable, efficient, automated processes for large-scale data analysis, machine-learning model development, model validation and serving. - Research new and innovative machine learning approaches. Why you will love this opportunity: Amazon is investing heavily in building a world-class advertising business. This team defines and delivers a collection of advertising products that drive discovery and sales. Our solutions generate billions in revenue and drive long-term growth for Amazon’s Retail and Marketplace businesses. We deliver billions of ad impressions, millions of clicks daily, and break fresh ground to create world-class products. We are a highly motivated, collaborative, and fun-loving team with an entrepreneurial spirit - with a broad mandate to experiment and innovate. Impact and Career Growth: You will invent new experiences and influence customer-facing shopping experiences to help suppliers grow their retail business and the auction dynamics that leverage native advertising; this is your opportunity to work within the fastest-growing businesses across all of Amazon! Define a long-term science vision for our advertising business, driven from our customers' needs, translating that direction into specific plans for research and applied scientists, as well as engineering and product teams. This role combines science leadership, organizational ability, technical strength, product focus, and business understanding. Team video https://youtu.be/zD_6Lzw8raE
US, WA, Bellevue
The Artificial General Intelligence (AGI) team is looking for a passionate, talented, and inventive Senior Applied Scientist with a strong deep learning background, to help build industry-leading technology with Large Language Models (LLMs) and multimodal systems. Key job responsibilities As a Senior Applied Scientist with the AGI team, you will work with talented peers to develop novel algorithms and modeling techniques to advance the state of the art with LLMs. Your work will directly impact our customers in the form of products and services that make use of audio technology. You will leverage Amazon’s heterogeneous data sources and large-scale computing resources to accelerate advances in AGI in audio domain. About the team Our team has a mission to push the envelope of AGI in audio domain, in order to provide the best-possible experience for our customers.
US, Virtual
Amazon is deeply invested in R&D with hundreds of researchers and applied scientists committed to innovation across every part of the company. The Amazon Scholars and Visiting Academic programs have broadened opportunities for academics to join Amazon in a flexible capacity, in particular part-time arrangements and sabbaticals. The program is designed for academics from universities around the globe who want to apply research methods in practice and help us solve hard technical challenges without leaving their academic institutions. We believe that Amazon is a unique place to measure the impact of new scientific ideas, given our scale and our ownership of both an information infrastructure and physical infrastructure. You will have a chance to have a ground-up impact on our systems, our business, and most importantly, our customers, through your expertise. Applications are accepted from academic experts in research areas including, but not limited to, the following: Artificial Intelligence, Avionics, Computer Vision, Data Science, Economics, Machine Learning, Optimization, Natural Language Processing, Quantum Computing, Robotics and Sustainability. Key job responsibilities As an Amazon Scholar or Visiting Academic, your responsibilities may include: * Advising business leaders on strategic plans * Diving deep to solve a specific technical problem in an organization’s roadmap * Advising junior researchers on methods.
KR, Seoul
Are you looking to work at the forefront of Machine Learning and AI? Would you be excited to apply cutting edge Generative AI algorithms to solve real world problems with significant impact? The Generative AI Innovation Center at AWS is a new strategic team that helps AWS customers implement Generative AI solutions and realize transformational business opportunities. This is a team of strategists, data scientists, engineers, and solution architects working step-by-step with customers to build bespoke solutions that harness the power of generative AI.' The team helps customers imagine and scope the use cases that will create the greatest value for their businesses, select and train and fine tune the right models, define paths to navigate technical or business challenges, develop proof-of-concepts, and make plans for launching solutions at scale. The GenAI Innovation Center team provides guidance on best practices for applying generative AI responsibly and cost efficiently. You will work directly with customers and innovate in a fast-paced organization that contributes to game-changing projects and technologies. You will design and run experiments, research new algorithms, and find new ways of optimizing risk, profitability, and customer experience. We’re looking for Data Scientists capable of using GenAI and other techniques to design, evangelize, and implement state-of-the-art solutions for never-before-solved problems. 머신 러닝과 AI의 최전선에서 일하고 싶으신가요? 최첨단 제너레이티브 AI 알고리즘을 적용하여 실제 문제를 큰 영향을 미치며 해결하게 되어 기대되시나요?AWS의 제너레이티브 AI 혁신 센터는 AWS 고객이 제너레이티브 AI 솔루션을 구현하고 혁신적인 비즈니스 기회를 실현할 수 있도록 지원하는 새로운 전략 팀입니다.전략가, 데이터 사이언티스트, 엔지니어 및 솔루션 아키텍트로 구성된 팀이 고객과 단계별로 협력하여 제너레이티브 AI의 힘을 활용하는 맞춤형 솔루션을 구축합니다. 이 팀은 고객 비즈니스에 최고의 가치를 창출할 사례를 구상 및 범위를 지정하고, 적합한 모델을 선택 및 교육 및 조정하고, 기술 또는 비즈니스 과제를 탐색하기 위한 경로를 정의하고, PoC을 개발하고, 대규모 솔루션 출시를 위한 계획을 세울 수 있도록 지원합니다.GenAI Innovation Center 팀은 제너레이티브 AI를 책임감 있고 비용 효율적으로 적용하는 모범 사례에 대한 지침을 제공합니다. 고객과 직접 협력하고 빠르게 변화하는 조직에서 혁신을 이루어 판도를 바꾸는 프로젝트 및 기술에 기여하게 될 것입니다.실험을 설계 및 실행하고, 새로운 알고리즘을 연구하고, 위험, 수익성 및 고객 경험을 최적화하는 새로운 방법을 찾게 됩니다. GenAI 및 기타 기술을 사용하여 이전에 해결되지 않은 문제에 대한 최첨단 솔루션을 설계, 홍보 및 구현할 수 있는 데이터 과학자를 찾고 있습니다. Key job responsibilities As a Data Scientist, you will: Collaborate with AI/ML scientists and architects to research, design, develop, and evaluate cutting-edge generative AI algorithms to address real-world challenges Interact with customers directly to understand the business problem, help and aid them in implementation of generative AI solutions, deliver briefing and deep dive sessions to customers and guide customer on adoption patterns and paths to production Create and deliver best practice recommendations, tutorials, blog posts, sample code, and presentations adapted to technical, business, and executive stakeholder Provide customer and market feedback to Product and Engineering teams to help define product direction 데이터 사이언티스트는 다음과 같은 일을 하게 됩니다. AI/ML 과학자 및 설계자와 협업하여 실제 문제를 해결하기 위한 최첨단 제너레이티브 AI 알고리즘을 연구, 설계, 개발 및 평가합니다. 고객과 직접 상호 작용하여 비즈니스 문제를 이해하고, 제너레이티브 AI 솔루션을 구현하도록 지원 및 지원하고, 고객에게 브리핑 및 심층 분석 세션을 제공하고, 도입 패턴 및 생산 경로를 안내합니다. 기술, 비즈니스 및 경영진 이해 관계자에게 적합한 모범 사례 권장 사항, 튜토리얼, 블로그 게시물, 샘플 코드 및 프레젠테이션을 만들어 제공합니다. 제품 및 엔지니어링 팀에 고객 및 시장 피드백을 제공하여 제품 방향을 정의하는 데 도움을 줍니다. About the team Sales, Marketing and Global Services (SMGS) AWS Sales, Marketing, and Global Services (SMGS) is responsible for driving revenue, adoption, and growth from the largest and fastest-growing small- and mid-market accounts to enterprise-level customers, including the public sector. The AWS Global Support team interacts with leading companies and believes that world-class support is critical to customer success. AWS Support also partners with a global list of customers that are building mission-critical applications on top of AWS services. The Professional Services team is part of Global Services. AWS 영업, 마케팅 및 글로벌 서비스 (SMGS) 는 크고 빠르게 성장하는, 공공 부문에서 엔터프라이즈에 이르기까지 고객의 성장을 돕는 역할을 합니다. AWS 글로벌 지원 팀은 글로벌 기업과 교류하며 고객의 성공을 돕습니다. 또한 AWS Support는 AWS 서비스를 기반으로 미션 크리티컬 애플리케이션을 구축하는 전 세계 고객 목록과도 파트너 관계를 맺고 있습니다. 프로페셔널 서비스팀은 AWS 내 글로벌 서비스팀에 소속되어 있습니다. About AWS Diverse Experiences AWS values diverse experiences. Even if you do not meet all of the qualifications and skills listed in the job description, we encourage candidates to apply. If your career is just starting, hasn’t followed a traditional path, or includes alternative experiences, don’t let it stop you from applying. Why AWS? Amazon Web Services (AWS) is the world’s most comprehensive and broadly adopted cloud platform. We pioneered cloud computing and never stopped innovating — that’s why customers from the most successful startups to Global 500 companies trust our robust suite of products and services to power their businesses. Inclusive Team Culture Here at AWS, it’s in our nature to learn and be curious. Our employee-led affinity groups foster a culture of inclusion that empower us to be proud of our differences. Ongoing events and learning experiences, including our Conversations on Race and Ethnicity (CORE) and AmazeCon (gender diversity) conferences, inspire us to never stop embracing our uniqueness. Mentorship & Career Growth We’re continuously raising our performance bar as we strive to become Earth’s Best Employer. That’s why you’ll find endless knowledge-sharing, mentorship and other career-advancing resources here to help you develop into a better-rounded professional. Work/Life Balance We value work-life harmony. Achieving success at work should never come at the expense of sacrifices at home, which is why we strive for flexibility as part of our working culture. When we feel supported in the workplace and at home, there’s nothing we can’t achieve in the cloud. 다양성 AWS는 다양한 경험을 중요하게 생각합니다. JD에 나와 있는 자격 및 기술을 모두 충족하지 못하더라도 지원자가 지원하도록 권장합니다. 경력을 이제 막 시작하였거나, 전통적인 경력을 쌓지 않았거나, 조금 다른 경험을 쌓았다고, 지원을 중단하실 필요는 없습니다. AWS를 선택해야 하는 이유 아마존 웹 서비스 (AWS) 는 세계에서 가장 포괄적이고 널리 채택된 클라우드 플랫폼입니다.우리는 클라우드 컴퓨팅 시장을 개척했으며 혁신을 멈추지 않았습니다. 이것이 바로 가장 성공적인 스타트업부터 Global 500 기업에 이르는 고객이 AWS의 제품 및 서비스 제품군을 신뢰하는 이유입니다. 일과 삶의 균형 우리는 일과 삶의 조화를 중요하게 생각합니다.직장에서의 성공을 위해 가정에서의 희생을 감수해서는 안 됩니다. 그렇기 때문에 유연한 근무 시간과 근무 방식이 우리 문화의 일부입니다.직장과 가정에서 지지받는다고 느낄 때 클라우드로는 달성할 수 없는 것이 없습니다. 포용적인 팀 문화 AWS에서는 배우고 호기심을 갖는 것이 우리의 본능입니다.직원이 주도하는 어피니티 그룹은 서로 다른 점을 자랑스럽게 여길 수 있는 포용의 문화를 조성합니다.인종 및 민족에 관한 대화 (CORE) 및 AmazeCon (성별 다양성) 컨퍼런스를 포함하여 진행 중인 이벤트와 학습 경험은 우리가 우리의 독창성을 받아들일 수 있도록 영감을 줍니다. 멘토십 및 경력 개발 우리는 세계 최고의 고용주가 되기 위해 노력하면서 지속적으로 성과 기준을 높이고 있습니다.그렇기 때문에 더 다재다능한 전문가로 발전하는 데 도움이 되는 지식 공유, 멘토십 및 기타 경력 개발 리소스를 찾을 수 있습니다. 일과 삶의 균형 우리는 일과 삶의 조화를 중요하게 생각합니다.직장에서의 성공을 위해 가정에서의 희생을 감수해서는 절대 안 됩니다. 이것이 바로 우리가 근무 문화의 일환으로 유연성을 추구하기 위해 노력하는 이유입니다.직장과 가정에서 지지받는다고 느낄 때 클라우드로는 달성할 수 없는 것이 없습니다. #aws-korea-proserv-ap #AWSKOREA
DE, Aachen
The Artificial General Intelligence (AGI) team is looking for a passionate, talented, and inventive Senior Applied Scientist with a strong deep learning background, to build industry-leading Generative Artificial Intelligence (GenAI) technology with Large Language Models (LLMs) and multimodal systems. Key job responsibilities As a Senior Applied Scientist with the AGI team, you will work with talented peers to lead the development of novel algorithms and modeling techniques, to advance the state of the art with LLMs. Your work will directly impact our customers in the form of products and services that make use of speech and language technology. You will leverage Amazon’s heterogeneous data sources and large-scale computing resources to accelerate advances in spoken language understanding. About the team The AGI team has a mission to push the envelope in GenAI with LLMs and multimodal systems, in order to provide the best-possible experience for our customers.
CA, BC, Vancouver
We are open to candidates located in: Seattle and Bellevue, Washington Atlanta, GA As a Senior Data Scientist, you will be on the ground floor with your team, shaping the way performance is measured, defining what questions should be asked, and scaling analytics methods and tools to support our growing business. You will work closely with Data Engineers, Product Managers, Business Intelligence Engineers, and Software Engineers to develop statistical models, design and run experiments, and find new ways to to optimize the customer experience. A successful candidate is highly analytical, able to work effectively in a matrix organization, and adept at synthesizing a variety of technologies and capabilities into products that enhances the PXF experience across multiple products. You must engage with customers to deeply understand their current and emerging needs. PXF applications are rapidly evolving and our user base is rapidly expanding, as a DS on the team you will own diving into the different users personas and inventing on behalf of the app users to meet their needs. We are looking for someone who's customer obsessed and technology savvy - with a passion for app development work. The ideal candidate will have a well-rounded technical background as well as a history of leading complex, ambiguous projects end-to-end. Key job responsibilities - Partner with business stakeholders in formulating the business problem and providing recommendations on the approach - Understanding customer behavior to personalize customer experience, build recommendation engines to provide relevant results to customers, customer lifecycle analysis and usage behavior - Conduct large scale A/B testing and offline/online experiments to evaluate performance of various programs and drive product improvements across partner teams - Process large scale datasets using distributed computing platform to build models, mining insights from data and prototyping models that optimize towards various business goals and metrics - Interact with cross-functional teams and make business recommendations i.e cost-benefit, forecasting, experiment analysis and present findings to leadership team - Driving product roadmap and prioritizations of science projects with the PMs to improve customer experience About the team PXF builds the employee experiences that connect Amazonians, support them through their employment journey, and make Amazon Earth's Best Employer. Our products include A to Z mobile application directly impacts the lives of associates by helping them identify the best shifts for their schedule, opportunities to pick up additional work, and choose when they get paid. We enable Amazon employees to easily find and access high-quality and authoritative content throughout their employment lifecycle through content management and Search capabilities. We also provide employees with a dynamic and ever-evolving learning experience to protect, prepare, and advance their careers.
IN, KA, Bengaluru
Advertising at Amazon is a fast-growing multi-billion dollar business that spans across desktop, mobile and connected devices; encompasses ads on Amazon and a vast network of hundreds of thousands of third party publishers; and extends across US, EU and an increasing number of international geographies. One of the key focus areas is Traffic Quality where we endeavour to identify non-human and invalid traffic within programmatic ad sources, and weed them out to ensure a high quality advertising marketplace. We do this by building machine learning and optimization algorithms that operate at scale, and leverage nuanced features about user, context, and creative engagement to determine the validity of traffic. The challenge is to stay one step ahead by investing in deep analytics and developing new algorithms that address emergent attack vectors in a structured and scalable fashion. We are committed to building a long-term traffic quality solution that encompasses all Amazon advertising channels and provides state-of-the-art traffic filtering that preserves advertiser trust and saves them hundreds of millions of dollars of wasted spend. We are looking for talented applied scientists who enjoy working on creative machine learning algorithms and thrive in a fast-paced, fun environment. An Applied Scientist is responsible for solving inherently hard problems in advertising fraud detection using deep learning, self-supervised techniques, representation learning and advanced clustering. An ideal candidate should have strong depth and breadth knowledge in machine learning, data mining and statistics. Traffic quality systems process billions of ad-impressions and clicks per day, by leveraging cutting-edge open source technologies like Hadoop, Spark, Redis and Amazon's cloud services like EC2, S3, EMR, DynamoDB and RedShift. The candidate should have reasonable programming and design skills to manipulate unstructured and big data and build prototypes that work on massive datasets. The candidate should be able to apply business knowledge to perform broad data analysis as a precursor to modeling and to provide valuable business intelligence.