# Matematički blogovi

### Adjoint Brascamp-Lieb inequalities

Jon Bennett and I have just uploaded to the arXiv our paper “Adjoint Brascamp-Lieb inequalities“. In this paper, we observe that the family of multilinear inequalities known as the Brascamp-Lieb inequalities (or Holder-Brascamp-Lieb inequalities) admit an adjoint formulation, and explore the theory of these adjoint inequalities and some of their consequences.

To motivate matters let us review the classical theory of adjoints for linear operators. If one has a bounded linear operator for some measure spaces and exponents , then one can define an adjoint linear operator involving the dual exponents , obeying (formally at least) the duality relation

for suitable test functions on respectively. Using the dual characterization of (and similarly for ), one can show that has the same operator norm as .There is a slightly different way to proceed using Hölder’s inequality. For sake of exposition let us make the simplifying assumption that (and hence also ) maps non-negative functions to non-negative functions, and ignore issues of convergence or division by zero in the formal calculations below. Then for any reasonable function on , we have

by (1) and Hölder; dividing out by we obtain , and a similar argument also recovers the reverse inequality.The first argument also extends to some extent to multilinear operators. For instance if one has a bounded bilinear operator for then one can then define adjoint bilinear operators and obeying the relations

and with exactly the same operator norm as . It is also possible, formally at least, to adapt the Hölder inequality argument to reach the same conclusion.In this paper we observe that the Hölder inequality argument can be modified in the case of Brascamp-Lieb inequalities to obtain a different type of adjoint inequality. (Continuous) Brascamp-Lieb inequalities take the form

for various exponents and surjective linear maps , where are arbitrary non-negative measurable functions and is the best constant for which this inequality holds for all such . [There is also another inequality involving variances with respect to log-concave distributions that is also due to Brascamp and Lieb, but it is not related to the inequalities discussed here.] Well known examples of such inequalities include Hölder’s inequality and the sharp Young convolution inequality; another is the Loomis-Whitney inequality, the first non-trivial example of which is for all non-negative measurable . There are also discrete analogues of these inequalities, in which the Euclidean spaces are replaced by discrete abelian groups, and the surjective linear maps are replaced by discrete homomorphisms.
The operation of pulling back a function on by a linear map to create a function on has an adjoint *pushforward map* , which takes a function on and basically integrates it on the fibers of to obtain a “marginal distribution” on (possibly multiplied by a normalizing determinant factor). The adjoint Brascamp-Lieb inequalities that we obtain take the form

*less*than or equal to ). For instance, the adjoints of the Loomis-Whitney inequality (2) are the inequalities for all non-negative measurable , all summing to , and all , where the exponents are defined by the formula and the are the marginals of :

One can derive these adjoint Brascamp-Lieb inequalities from their forward counterparts by a version of the Hölder inequality argument mentioned previously, in conjunction with the observation that the pushforward maps are mass-preserving (i.e., they preserve the norm on non-negative functions). Conversely, it turns out that the adjoint Brascamp-Lieb inequalities are only available when the forward Brascamp-Lieb inequalities are. In the discrete case the forward and adjoint Brascamp-Lieb constants are essentially identical, but in the continuous case they can (and often do) differ by up to a constant. Furthermore, whereas in the forward case there is a famous theorem of Lieb that asserts that the Brascamp-Lieb constants can be computed by optimizing over gaussian inputs, the same statement is only true up to constants in the adjoint case, and in fact in most cases the gaussians will fail to optimize the adjoint inequality. The situation appears to be complicated; roughly speaking, the adjoint inequalities only use a portion of the range of possible inputs of the forward Brascamp-Lieb inequality, and this portion often misses the gaussian inputs that would otherwise optimize the inequality.

We have located a modest number of applications of the adjoint Brascamp-Lieb inequality (but hope that there will be more in the future):

- The inequalities become equalities at ; taking a derivative at this value (in the spirit of the replica trick in physics) we recover the entropic Brascamp-Lieb inequalities of Carlen and Cordero-Erausquin. For instance, the derivative of the adjoint Loomis-Whitney inequalities at yields Shearer’s inequality.
- The adjoint Loomis-Whitney inequalities, together with a few more applications of Hölder’s inequality, implies the log-concavity of the Gowers uniformity norms on non-negative functions, which was previously observed by Shkredov and by Manners.
- Averaging the adjoint Loomis-Whitney inequalities over coordinate systems gives reverse inequalities for the X-ray transform and other tomographic transforms that appear to be new in the literature. In particular, we obtain some monotonicity of the norms or entropies of the -plane transform in (if the exponents are chosen in a dimensionally consistent fashion).

We also record a number of various of the adjoint Brascamp-Lieb inequalities, including discrete variants, and a reverse inequality involving norms with rather than .

### Sumsets and entropy revisited

Ben Green, Freddie Manners and I have just uploaded to the arXiv our preprint “Sumsets and entropy revisited“. This paper uses entropy methods to attack the Polynomial Freiman-Ruzsa (PFR) conjecture, which we study in the following two forms:

**Conjecture 1 (Weak PFR over )** Let be a finite non-empty set whose doubling constant is at most . Then there is a subset of of density that has affine dimension (i.e., it is contained in an affine space of dimension ).

**Conjecture 2 (PFR over )** Let be a non-empty set whose doubling constant is at most . Then can be covered by cosets of a subspace of cardinality at most .

Our main results are then as follows.

**Theorem 3** If with , then

- (i) There is a subset of of density of “skew-dimension” (or “query complexity”) .
- (ii) There is a subset of of density of affine dimension (where goes to zero as ).
- (iii) If Conjecture 2 holds, then there is a subset of of density of affine dimension . In other words, Conjecture 2 implies Conjecture 1.

The skew-dimension of a set is a quantity smaller than the affine dimension which is defined recursively; the precise definition is given in the paper, but suffice to say that singleton sets have dimension , and a set whose projection to has skew-dimension at most , and whose fibers in have skew-dimension at most for any , will have skew-dimension at most . (In fact, the skew-dimension is basically the largest quantity which obeys all of these properties.)

Part (i) of this theorem was implicitly proven by Pálvölgi and Zhelezov by a different method. Part (ii) with replaced by was established by Manners. To our knowledge, part (iii) is completely new.

Our proof strategy is to establish these combinatorial additive combinatorics results by using entropic additive combinatorics, in which we replace sets with random variables , and cardinality with (the exponential of) Shannon entropy. This is in order to take advantage of some superior features of entropic additive combinatorics, most notably good behavior with respect to homomorphisms.

For instance, the analogue of the combinatorial doubling constant of a finite non-empty subset of an abelian group , is the entropy doubling constant

of a finitely-valued random variable in , where are independent copies of and denotes Shannon entropy. There is also an analogue of the Ruzsa distance between two finite non-empty subsets of , namely the entropic Ruzsa distance where are independent copies of respectively. (Actually, one thing we show in our paper is that the independence hypothesis can be dropped, and this only affects the entropic Ruzsa distance by a factor of three at worst.) Many of the results about sumsets and Ruzsa distance have entropic analogues, but the entropic versions are slightly better behaved; for instance, we have a contraction property whenever is a homomorphism. In fact we have a refinement of this inequality in which the gap between these two quantities can be used to control the entropic distance between “fibers” of (in which one conditions and to be fixed). On the other hand, there are direct connections between the combinatorial and entropic sumset quantities. For instance, if is a random variable drawn uniformly from , then Thus if has small doubling, then has small entropic doubling. In the converse direction, if has small entropic doubling, then is close (in entropic Ruzsa distance) to a uniform random variable drawn from a set of small doubling; a version of this statement was proven in an old paper of myself, but we establish here a quantitatively efficient version, established by rewriting the entropic Ruzsa distance in terms of certain Kullback-Liebler divergences.Our first main result is a “99% inverse theorem” for entropic Ruzsa distance: if is sufficiently small, then there exists a finite subgroup of such that

This result uses the results just mentioned to relate to a set of small doubling, which can then be related to a subgroup by standard inverse theorems; this gives a weak version of (1) (roughly speaking losing a square root in the bound), and some additional analysis is needed to bootstrap this initial estimate back to (1).We now sketch how these tools are used to prove our main theorem. For (i), we reduce matters to establishing the following bilinear entropic analogue: given two non-empty finite subsets of , one can find subsets , with

such that have skew-dimension at most , for some absolute constant . This can be shown by an induction on (say). One applies a non-trivial coordinate projection to . If and are very close in entropic Ruzsa distance, then the 99% inverse theorem shows that these random variables must each concentrate at a point (because has no non-trivial finite subgroups), and can pass to a fiber of these points and use the induction hypothesis. If instead and are far apart, then by the behavior of entropy under projections one can show that the fibers of and under are closer on average in entropic Ruzsa distance of and themselves, and one can again proceed using the induction hypothesis.For parts (ii) and (iii), we first use an entropic version of an observation of Manners that sets of small doubling in must be irregularly distributed modulo . A clean formulation of this in entropic language is the inequality

whenever take values in a torsion-free abelian group such as ; this turns out to follow from two applications of the entropy submodularity inequality. One corollary of this (and the behavior of entropy under projections) is that This is the key link between the and worlds that is used to prove (ii), (iii): while (iii) relies on the still unproven PFR conjecture over , (ii) uses the unconditional progress on PFR by Konyagin, as detailed in this survey of Sanders. The argument has a similar inductive structure to that used to establish (i) (and if one is willing to replace by then the argument is in fact relatively straightforward and does not need any deep partial results on the PFR).As one byproduct of our analysis we also obtain an appealing entropic reformulation of Conjecture 2, namely that if is an -valued random variable then there exists a subspace of such that

Right now the best result in this direction is for any , by using Konyagin’s partial result towards the PFR.

### Salem Prize now accepting nominations

The Salem prize was established in 1968 and named in honor of Raphaël Salem (1898-1963), a mathematician famous notably for his deep study of the links between Fourier series and number theory and for pioneering applications of probabilistic methods to these fields. It was largely administered by Jean Bourgain until his untimely death in 2018; due to this and the COVID-19 pandemic, no prize was awarded for the years of 2019-2022. A list of past winners may be found here.

However, I am happy to report that the prize has been reactivated, and is now formally hosted by the Institute for Advanced Study, with Akshay Venkatesh overseeing the administration of the prize, and is accepting nominations for the 2023 Salem Prize until September 1st. Nominations should include a CV of the nominee and a nomination letter explaining the significance of the nominee’s work. Supplementary documentation, such as supporting letters of recommendation or key publications, can additionally be provided, but are not required.

Nominees may be individuals from any country or institution. Preference will be given to nominees who have received their PhD in the last ten years, although this rule may be relaxed if there are mitigating personal circumstances, or if there have been few Salem prize winners in recent years. Self-nominations will not be considered, nor are past Prize winners or Scientific Committee members eligible.

The prize does not come with a direct monetary award, but winners will be invited to visit the IAS and to give a lecture associated with the award of the prize.

I will be chairing the Scientific Committee to evaluate the nominations and recommend a prize winner; Guy David and Mikhail Sodin have also agreed to serve. Special thanks also to Peter Sarnak for his tireless efforts to ensure the continuation of the prize after Jean’s passing.

### Sums of GUE matrices and concentration of hives from correlation decay of eigengaps

Hariharan Narayanan, Scott Sheffield, and I have just uploaded to the arXiv our paper “Sums of GUE matrices and concentration of hives from correlation decay of eigengaps“. This is a personally satisfying paper for me, as it connects the work I did as a graduate student (with Allen Knutson and Chris Woodward) on sums of Hermitian matrices, with more recent work I did (with Van Vu) on random matrix theory, as well as several other results by other authors scattered across various mathematical subfields.

Suppose are two Hermitian matrices with eigenvalues and respectively (arranged in non-increasing order. What can one say about the eigenvalues of the sum ? There are now many ways to answer this question precisely; one of them, introduced by Allen and myself many years ago, is that there exists a certain triangular array of numbers called a “hive” that has as its boundary values. On the other hand, by the pioneering work of Voiculescu in free probability, we know in the large limit that if are asymptotically drawn from some limiting distribution, and and are drawn independently at random (using the unitarily invariant Haar measure) amongst all Hermitian matrices with the indicated eigenvalues, then (under mild hypotheses on the distribution, and under suitable normalization), will almost surely have a limiting distribution that is the free convolution of the two original distributions.

One of my favourite open problems is to come up with a theory of “free hives” that allows one to explain the latter fact from the former. This is still unresolved, but we are now beginning to make a bit of progress towards this goal. We know (for instance from the calculations of Coquereaux and Zuber) that if are drawn independently at random with eigenvalues , then the eigenvalues of are distributed according to the boundary values of an “augmented hive” with two boundaries , drawn uniformly at random from the polytope of all such augmented hives. (This augmented hive is basically a regular hive with another type of pattern, namely a *Gelfand-Tsetlin pattern*, glued to one side of it.) So, if one could show some sort of concentration of measure for the entries of this augmented hive, and calculate what these entries concentrated to, one should presumably be able to recover Voiculescu’s result after some calculation.

In this paper, we are able to accomplish the first half of this goal, assuming that the spectra are not deterministic, but rather drawn from the spectra of rescaled GUE matrices (thus are independent rescaled copies of the GUE ensemble). We have chosen to normalize matters so that the eigenvalues have size , so that the entries of the augmented hive have entries . Our result is then that the entries of the augmented hive in fact have a standard deviation of , thus exhibiting a little bit of concentration. (Actually, from the Brunn-Minkowski inequality, the distribution of these entries is log concave, so once once controls the standard deviation one also gets a bit of exponential decay beyond the standard deviation; Narayanan and Sheffield had also recently established the existence of a rate function for this sort of model.) Presumably one should get much better concentration, and one should be able to handle other models than the GUE ensemble, but this is the first advance that we were able to achieve.

Augmented hives seem tricky to work with directly, but by adapting the octahedron recurrence introduced for this problem by Knutson, Woodward, and myself some time ago (which is related to the associativity of addition for Hermitian matrices), one can construct a piecewise linear volume-preserving map between the cone of augmented hives, and the product of two Gelfand-Tsetlin cones. The problem then reduces to establishing concentration of measure for certain piecewise linear maps on products of Gelfand-Tsetlin cones (endowed with a certain GUE-type measure). This is a promising formulation because Gelfand-Tsetlin cones are by now quite well understood.

On the other hand, the piecewise linear map, initially defined by iterating the octahedron relation , looks somewhat daunting. Fortunately, there is an explicit formulation of this map due to Speyer, as the supremum of certain linear maps associated to perfect matchings of a certain “excavation graph”. For us it was convenient to work with the dual of this excavation graph, and associate these linear maps to certain “lozenge tilings” of a hexagon.

It would be more convenient to study the concentration of each linear map separately, rather than their supremum. By the Cheeger inequality, it turns out that one can relate the latter to the former provided that one has good control on the Cheeger constant of the underlying measure on the Gelfand-Tsetlin cones. Fortunately, the measure is log-concave, so one can use the very recent work of Klartag on the KLS conjecture to eliminate the supremum (up to a logarithmic loss which is only moderately annoying to deal with).

It remains to obtain concentration on the linear map associated to a given lozenge tiling. After stripping away some contributions coming from lozenges near the edge (using some eigenvalue rigidity results of Van Vu and myself), one is left with some bulk contributions which ultimately involve eigenvalue interlacing gaps such as

where is the eigenvalue of the top left minor of , and is in the bulk region for some fixed . To get the desired result, one needs some non-trivial correlation decay in for these statistics. If one was working with eigenvalue gaps rather than interlacing results, then such correlation decay was conveniently obtained for us by recent work of Cippoloni, Erdös, and Schröder. So the last remaining challenge is to understand the relation between eigenvalue gaps and interlacing gaps.For this we turned to the work of Metcalfe, who uncovered a determinantal process structure to this problem, with a kernel associated to Lagrange interpolation polynomials. It is possible to satisfactorily estimate various integrals of these kernels using the residue theorem and eigenvalue rigidity estimates, thus completing the required analysis.

### AI Anthology

Back in March, I was approached to contribute to a then-upcoming anthology project to evaluate an early access version of the GPT-4 large language model, and write a short essay about my experiences. Our prompt was to focus on two core questions:

- How might this technology and its successors contribute to human flourishing?
- How might we as society best guide the technology to achieve maximal benefits for humanity?

The anthology is now in the process of being rolled out, with twelve of the twenty essays, including mine, public at this time of writing.

As an experiment, I also asked GPT-4 itself to contribute an essay to the anthology from the same prompts (and playing the role of a research mathematician), then I gave it my own essay (which I wrote independently) and asked it both to rewrite its own essay in the style of my own, or to copyedit my essay into what it deemed to be a better form. I recorded the results of those experiments here; the output was reasonably well written and on topic, but not exceptional in content.

### A note on the mean value of the Hooley delta function

Dimitris Koukoulopoulos and I have just uploaded to the arXiv our paper “A note on the mean value of the Hooley delta function“. This paper concerns a (still somewhat poorly understood) basic arithmetic function in multiplicative number theory, namely the *Hooley delta function*

On the other hand, the statistical behavior of the Hooley delta function is significantly less well understood, even conjecturally. Again drawing at random from to for large , the median is known to be somewhere between and for large – a (difficult) recent result of Ford, Green, and Koukoulopolous (for the lower bound) and La Bretèche and Tenenbaum (for the upper bound). And the mean was even less well controlled; the best previous bounds were

for any , with the lower bound due to Hall and Tenenbaum, and the upper bound a recent result of La Bretèche and Tenenbaum.The main result of this paper is an improvement of the upper bound to

It is still unclear to us exactly what to conjecture regarding the actual order of the mean value.The reason we looked into this problem was that it was connected to forthcoming work of David Conlon, Jacob Fox, and Huy Pham on the following problem of Erdos: what is the size of the largest subset of with the property that no non-empty subset of sums to a perfect square? Erdos observed that one can obtain sets of size (basically by considering certain homogeneous arithmetic progressions), and Nguyen and Vu showed an upper bound of . With our mean value bound as input, together with several new arguments, Conlon, Fox, and Pham have been able to improve the upper bound to .

Let me now discuss some of the ingredients of the proof. The first few steps are standard. Firstly we may restrict attention to square-free numbers without much difficulty (the point being that if a number factors as with squarefree, then ). Next, because a square-free number can be uniquely factored as where is a prime and lies in the finite set of squarefree numbers whose prime factors are less than , and , it is not difficult to establish the bound

The upshot of this is that one can replace an ordinary average with a logarithmic average, thus it suffices to show We actually prove a slightly more refined distributional estimate: for any , we have a bound outside of an exceptional set which is small in the sense that It is not difficult to get from this distributional estimate to the logarithmic average estimate (1) (worsening the exponent to ).To get some intuition on the size of , we observe that if and is the factor of coming from the prime factors less than , then

On the other hand, standard estimates let one establish that for all , and all outside of an exceptional set that is small in the sense (3); in fact it turns out that one can also get an additional gain in this estimate unless is close to , which turns out to be useful when optimizing the bounds. So we would like to approximately reverse the inequalities in (4) and get from (5) to (2), possibly after throwing away further exceptional sets of size (3).At this point we perform another standard technique, namely the moment method of controlling the supremum by the moments

for natural numbers ; it is not difficult to establish the bound and one expects this bound to become essentially sharp once . We will be able to show a moment bound for any for some exceptional set obeying the smallness condition (3) (actually, for technical reasons we need to improve the right-hand side slightly to close an induction on ); this will imply the distributional bound (2) from a standard Markov inequality argument (setting ).The strategy is then to obtain a good recursive inequality for (averages of) . As in the reduction to (1), we factor where is a prime and . One observes the identity

for any ; taking moments, one obtains the identity As in previous literature, one can try to average in here and apply Hölder’s inequality. But it convenient to first use the symmetry of the summand in to reduce to the case of relatively small values of : One can extract out the term as It is convenient to eliminate the factor of by dividing out by the divisor function: This inequality is suitable for iterating and also averaging in and . After some standard manipulations (using the Brun–Titchmarsh and Hölder inequalities), one is able to estimate sums such as in terms of sums such as (assuming a certain monotonicity property of the exceptional set that turns out to hold in our application). By an induction hypothesis and a Markov inequality argument, one can get a reasonable pointwise upper bound on (after removing another exceptional set), and the net result is that one can basically control the sum (6) in terms of expressions such as for various . This allows one to estimate these expressions efficiently by induction.

### Inequalities, convergence, and continuity as “special deals”

The “epsilon-delta” nature of analysis can be daunting and unintuitive to students, as the heavy reliance on inequalities rather than equalities. But it occurred to me recently that one might be able to leverage the intuition one already has from “deals” – of the type one often sees advertised by corporations – to get at least some informal understanding of these concepts.

Take for instance the concept of an upper bound or a lower bound on some quantity . From an economic perspective, one could think of the upper bound as an assertion that can be “bought” for units of currency, and the lower bound can similarly be viewed as an assertion that can be “sold” for units of currency. Thus for instance, a system of inequalities and equations like

could be viewed as analogous to a currency rate exchange board, of the type one sees for instance in airports:Currency We buy at We sell at – –

Someone with an eye for spotting “deals” might now realize that one can actually buy for units of currency rather than , by purchasing one copy each of and for units of currency, then selling off to recover units of currency back. In more traditional mathematical language, one can improve the upper bound to by taking the appropriate linear combination of the inequalities , , and . More generally, this way of thinking is useful when faced with a linear programming situation (and of course linear programming is a key foundation for operations research), although this analogy begins to break down when one wants to use inequalities in a more non-linear fashion.

Asymptotic estimates such as (also often written or ) can be viewed as some sort of liquid market in which can be used to purchase , though depending on market rates, one may need a large number of units of in order to buy a single unit of . An asymptotic estimate like represents an economic situation in which is so much more highly desired than that, if one is a patient enough haggler, one can eventually convince someone to give up a unit of for even just a tiny amount of .

When it comes to the basic analysis concepts of convergence and continuity, one can similarly view these concepts as various economic transactions involving the buying and selling of *accuracy*. One could for instance imagine the following hypothetical range of products in which one would need to spend more money to obtain higher accuracy to measure weight in grams:

Object Accuracy Price Low-end kitchen scale gram High-end bathroom scale grams Low-end lab scale grams High-end lab scale grams

The concept of convergence of a sequence to a limit could then be viewed as somewhat analogous to a rewards program, of the type offered for instance by airlines, in which various tiers of perks are offered when one hits a certain level of “currency” (e.g., frequent flyer miles). For instance, the convergence of the sequence to its limit offers the following accuracy “perks” depending on one’s level in the sequence:

Status Accuracy benefit Eligibility Basic status Bronze status Silver status Gold status

With this conceptual model, convergence means that any status level of accuracy can be unlocked if one’s number of “points earned” is high enough.

In a similar vein, continuity becomes analogous to a conversion program, in which accuracy benefits from one company can be traded in for new accuracy benefits in another company. For instance, the continuity of the function at the point can be viewed in terms of the following conversion chart:

Accuracy benefit of to trade in Accuracy benefit of obtained

Again, the point is that one can purchase any desired level of accuracy of provided one trades in a suitably high level of accuracy of .

At present, the above conversion chart is only available at the single location . The concept of uniform continuity can then be viewed as an advertising copy that “offer prices are valid in all store locations”. In a similar vein, the concept of equicontinuity for a class of functions is a guarantee that “offer applies to all functions in the class , without any price discrimination. The combined notion of *uniform equicontinuity* is then of course the claim that the offer is valid in all locations and for all functions.

In a similar vein, differentiability can be viewed as a deal in which one can trade in accuracy of the input for approximately linear behavior of the output; to oversimplify slightly, smoothness can similarly be viewed as a deal in which one trades in accuracy of the input for high-accuracy polynomial approximability of the output. Measurability of a set or function can be viewed as a deal in which one trades in a level of resolution for an accurate approximation of that set or function at the given resolution. And so forth.

Perhaps readers can propose some other examples of mathematical concepts being re-interpreted as some sort of economic transaction?

### AI to Assist Mathematical Reasoning: A Workshop

The National Academies of Science, Engineering, and Mathematics are hosting a virtual workshop on the topic of “AI to Assist Mathematical Reasoning” from June 12-14. The tentative program can be found here. I am one of the members of the organizing committee for this workshop, together with Petros Koumoutsakos, Jordan Ellenberg, Melvin Greer, Brendan Hassett, Yann A. LeCun, Heather Macbeth, Talia Ringer, Kavitha Srinivas, and Michelle Schwalbe. There is some thematic overlap (and a few speakers in common) with the recent IPAM program on machine assisted proof, though with more of a focus on the current and projected technical capabilities of machine learning algorithms for mathematics. Registration for the event is currently open at the web page for the workshop.

### PCAST Working Group on Generative AI Invites Public Input

As part of my duties on the Presidential Council of Advisors on Science and Technology (PCAST), I am co-chairing (with Laura Greene) a working group studying the impacts of generative artificial intelligence technology (which includes popular text-based large language models such as ChatGPT or diffusion model image generators such as DALL-E 2 or Midjourney, as well as models for scientific applications such as protein design or weather prediction), both in science and in society more broadly. To this end, we will have public sessions on these topics during our PCAST meeting next week on Friday, May 19, with presentations by the following speakers, followed by an extensive Q&A session:

- AI enabling science:
- Anima Anandkumar (Caltech & NVIDIA)
- Demis Hassabis (Deepmind)
- Fei-Fei Li (Stanford)

- AI and society:
- Sendhil Mullainathan (Chicago)
- Daron Acemoglu (MIT)
- Sarah Kreps (Cornell)

The event will be livestreamed on the PCAST meeting page. I am personally very much looking forward to these sessions, as I believe they will be of broad public interest.

In parallel to this, our working group is also soliciting public input for submissions from the public on how to identify and promote the beneficial deployment of generative AI, and on how best to mitigate risks. Our initial focus is on the challenging topic of how to detect, counteract, and mitigate AI-generated disinformation and “deepfakes”, without sacrificing the freedom of speech and public engagement with elected officials that is needed for a healthy democracy to function; in the future we may also issue further requests centered around other aspects of generative AI. Further details of our request, and how to prepare a submission, can be found at this link.

We also encourage submissions to some additional requests for input on AI-related topics by other agencies:

- The Office of Science Technology and Policy (OSTP) Request for Information on how automated tools are being used to surveil, monitor, and manage workers.
- The National Telecommunications and Information Administration (NTIA) request for comment on AI accountability policy.

Readers who wish to know more about existing or ongoing federal AI policy efforts may also be interested in the following resources:

- The White House Blueprint for an AI Bill of Rights lays out core aspirational principles to guide the responsible design and deployment of AI technologies.
- The National Institute of Standards and Technology (NIST) released the AI Risk Management Framework to help organizations and individuals characterize and manage the potential risks of AI technologies.
- Congress created the National Security Commission on AI, which studied opportunities and risks ahead and the importance of guiding the development of AI in accordance with American values around democracy and civil liberties.
- The National Artificial Intelligence Initiative was launched to ensure U.S. leadership in the responsible development and deployment of trustworthy AI and support coordination of U.S. research, development, and demonstration of AI technologies across the Federal government.
- In January 2023, the Congressionally mandated National AI Research Resource (NAIRR) Task Force released an implementation plan for providing computational, data, testbed, and software resources to AI researchers affiliated with U.S organizations.

### Elias M. Stein Prize for New Perspectives in Analysis

The Elias M. Stein Prize for New Perspectives in Analysis is awarded for the development of groundbreaking methods in analysis which demonstrate promise to revitalize established areas or create new opportunities for mathematical discovery. The current prize amount is US$5,000 and the prize is awarded every three years for work published in the preceding six years.

This prize was endowed in 2022 by students, colleagues, and friends of Elias M. Stein (my former advisor) to honor his remarkable legacy in the area of mathematical analysis. Stein, who passed away in 2018, is remembered for identifying many deep principles and methods which transcend their original context, and for opening entirely new areas of research which captivated the attention and imagination of generations of analysts. This prize seeks to recognize mathematicians at any career stage who, like Stein, have found exciting new avenues for mathematical exploration in subjects old or new or made deep insights which demonstrate promise to reshape thinking across areas.

This will be the inaugural year for the prize, and I have agreed to serve on the prize committee. We welcome nominations for the prize, which will be accepted until June 30, 2023, and are seeking a strong and diverse pool of nominees. Nominations (submitted at this link) should include a letter of nomination and a brief citation to be used in the event that the nomination is successful. Alternatively, if you are aware of a strong potential candidate but are not able to provide the nomination yourself, we welcome your suggestion (by private email) along with — if possible — your suggestions of possible nominators.

For questions about this award, please contact the AMS Secretary at secretary@ams.org.