SPHEREx workshop, day 2

I got up at 0530 and looked at the participants and schedule for the SPHEREx workshop. I realized that I had prepared precisely the wrong talk yesterday! So I threw away my slides and made completely new slides. It was rushed. I forgot things. But it was still an improvement. I switched from saying things about scientific goals to saying things about technical improvements or extensions that could make the project more capable in respects that would serve the needs of (among other things) stellar science.

I then headed in to the workshop; I could only make it to the second day. I learned so much today. I can't do it justice. Here are some random facts: A lot could be learned about exoplanets if we could get bolometric fluxes for the stars.
I knew this already, I guess, but the prospects for SPHEREx here are excellent, if the project can deliver absolutely calibrated flux densities. There is a mass–metallicity relationship inside the Solar System! The Solar System contains Trojan satellites/asteroids around Neptune, not just Jupiter! There is no model for the zodiacal light in the Solar System that matches the observations to the level of precision that an infrared survey would need to remove or avoid it. The zodiacal light is consistent with being made up of ground up asteroids and evaporated comets! ALMA has observed many debris disks around nearby stars; some of these are angularly huge. The poster child is Fomalhaut, which has a thin, elliptical ring. It's a crazy thing. I learned these things from a combination of Dan Stevens (OSU), Jennifer Burt (MIT), Carey Lisse (JHU), and Meredith MacGregor (Harvard), but that's just a tiny sampling.

At the end of the day there was discussion of calibration, led by Doug Finkbeiner (CfA) and me. I very much enjoy the technical challenges for SPHEREx and the enthusiasm of the team taking them on.


slides prep

It was a very low-research day! But on the train to Boston, I prepared slides for a short talk at a meeting at Harvard about the SPHEREx mission concept. I wrote about how this cosmology mission (line intensity mapping and large-scale structure) might revolutionize our knowledge of stars in the Milky Way.


asteroseismic binaries; distances between transients

Simon J Murphy (Sydney) is in town for two weeks of hacking with Dan Foreman-Mackey (Flatiron). On arrival last week, the two of them implemented something I have been wanting to do for a long time, which is use asteroseismic phase shifts to find binary companions (yes, people have done this for a while now) but without binning the data up or ever explicitly measuring any time delays in bins or at times. This week (having solved that) they are looking at radial-velocity predictions from those discoveries, and testing them with HIRES spectra. They teamed up with Megan Bedell (Flatiron) to use her wobble system to make these measurements. All I did was cheer-lead.

In the afternoon, Alex Malz (NYU) and I discussed what we might do in an upcoming LSST transient classification challenge. I am interested in the following question: Say you have two sparsely and irregularly sampled light-curves of two transient events that are intrinsically similar but maybe at different redshifts, and you want to see that they are similar. How do you construct a relevant, useful, and tractable similarity or distance metric? I have lots of ideas; if we can solve this, we might have something to contribute.



In a low-research day, Megan Bedell and I went through the assumptions underlying our beliefs about chemical enrichment of a star from infall by dust or rocky material. This is relevant because she is finishing up a paper on the subject, with her extremely precise measurements of Solar twins.


EPRV, locked planets, data sharing

At lunch-time today, Megan Bedell (Flatiron) and Ray Pierrehumbert (Oxford) gave talks at Flatiron. During Bedell's talk, she nicely laid out the large number of results we have on extreme-precision radial-velocity measurement; we need to start writing papers asap! She even gave a very simple and new description of what we found with respect to HARPS wavelength-calibration fidelity, a couple of years ago. So we need to write that up too.

Pierrehumbert showed fluid-dynamics results on atmospheres of tidally-locked planets (which are interesting, because they sustain huge temperature gradients around their surfaces). He has some cases where he can't find any steady-state solution for the atmophere; the resulting time dependences might have observable consequences.

Late in the day, I gave a presentation to the AAAC that oversees astrophysics and inter-agency cooperation in astronomy across NSF, NASA, and DOE. I was asked to speak about the future of data sharing, data re-use, and joint analyses. I drew inspiration from cosmology and went into two of my standard sets of talking points: The first is that we need to be thinking about likelihood functions, and how to share them: Data sets are combined by their (possibly partially marginalized) likelihood functions. The second is that when data get sophisticated or complex, there is no point in releasing it without also releasing the code that made sense of the data in real scientific projects. That is, code and data releases can't really be seen as separate things. And we might not be able to have a data release without having a code release (with appropriate licensing for repurposing and re-use). My slides were incomplete, but I put them up here anyway.


get humans out of the loop for EPRV

At breakfast I went off on my complaint that if you have humans integrated into the operational decision-making of any astronomical project, you can destroy the statistical or legacy value of your data. This is a complaint I have had about some extreme-precision radial-velocity projects: Some investigators make decisions, before or during an observing run, to maximize planet yield that involve people in a room, talking. And they are talking about the outcomes of previous observations. That leads to sample selections that depend on the data in complex and un-model-able ways. As in: You would need a complete simulation of the human investigators and their group decision-making even to simulate the observing process, let alone build a probabilistic model of it. That's been a disaster for radial-velocity experiments, and one of the reasons that most of the populations inferences for exoplanets have come from Kepler. Of course, in their defense, many of the radial-velocity projects were trying to maximize planet yield, with statistics be damned!

This made me realize: We should look at things in the experimental-design and active-learning fields to see if we could obtain all the planet-yield benefits that exoplanet projects get from adaptively changing their observing mid-stream, but retain the usefulness of the data for populations inferences, by going to simple algorithmic replacements for the proverbial smoke-filled room. My intuition is a big yes. Another intuition is that any adaptive observing algorithm will necessarily be tuned to a science goal, so you will have to decide in detail what you value in addition to planet yield.

At Stars Group Meeting, John Brewer (Yale) described in detail the EXPRES radial-velocity spectrograph that is being built and commissioned for the Discovery Channel Telescope. He described design decisions, with many valuable contextual comments. One of these is that gas-cell spectrograph designs seem to be declining in popularity, while rigidly controlled lab-bench designs (so calibration parameters vary little and slowly) are rising in popularity. There are many radial-velocity projects getting started right now, so the landscape for this kind of work is changing fast. (Hence the above comments about experimental design!)


galaxies and dark matter

In another low-research day (it is beginning of term here) there was a great Astrophysics Seminar at NYU by Marilena Loverde (Stony Brook). She has the great capability of turning complex ideas, that are backed by cosmological simulations, into very simple conceptual arguments. And she has been doing this well for a decade now! She showed us that the bias—or the relationship between galaxies and the dark-matter field—cannot be a function of only local density (on some scale). It must also depend on the local expansion history, which depends on perturbations in the dark-matter and dark-energy-ish (quintessence) and radiation fields. This all gives me great hope for my anomalies project with Kate Storey-Fisher (NYU).


ready to resubmit; new latent-variable model

Combined with effort over the weekend, today I finished my revisions (in response to referee) for the MCMC paper I have written with Dan Foreman-Mackey (Flatiron). The paper is on arXiv but we will update it to the new version if it gets accepted. It is a relief to get it done. And the referee comments were very constructive and valuable. What amazes me is that the AAS Journals are willing to publish such an odd paper. I appreciate it, though: The AAS Journals are great journals.

I got a moment in with Christina Eilers (MPIA) to propose (yet another) latent-variable model for stellar spectra. The idea is that stellar spectra are very simple, so the relationship between the latent variables and the spectra could be linear! The relationship between the spectra and the parameters of interest is definitely non-linear, so we use a Gaussian Process to model the relationship between the latents and the labels. It is a hybrid linear–GP latent-variable model, or HLGPLVM! Oh now that's a great name.



A low-research day was saved by a great talk by Blakesley Burkhart (CfA) about turbulence in astrophysical MHD. She made an unassailable argument that if we don't understand turbulence (and we don't), we don't understand almost any astrophysical phenomenon. That is only slightly an exaggeration! And then she talked about a general framework for understanding turbulence: Simulate the hell out of it, build empirical statistics from those simulations, and use those statistics to measure (latent) physical parameters in observed systems (like interstellar molecular clouds and supernova remnants). She showed baby steps towards this ultimate formalism, but, in my way of thinking, this approach asymptotes to something like ABC or likelihood-free inference, in which we try to put posterior constraints on parameters of interest using statistical surrogates to obviate an explicit likelihood function. Right now, this seems like the only approach for something as complicated as turbulent plasmas.


inferring stellar parameters from light curves

Today at Flatiron, Melissa Ness (Flatiron) showed Dan Foreman-Mackey (Flatiron) and me that she can infer stellar effective temperature and surface gravity (but not metallicity!) from a Kepler light-curve, with a data-driven model (supervised regression). She feature-izes the light-curve into something for the model by taking the auto-correlation function. This is a clever idea, because it removes phase, or time-translation effects, from the data. And in a cross-validation she can show that she can infer temperatures to something like the precision with which they are measured in the training set! So this is extremely promising, and an interesting extension of things we have seen previously with stellar jitter. I opined that this could only get better if she looks into feature engineering; the autocorrelation function on a uniform grid cannot be the best feature set in any sense. But with encouragement from Foreman-Mackey, we decided to move ahead with a project now and leave feature engineering to later work.


Gaia and exoplanets

At this morning's Gaia DR2 prep workshop (parallel-working meeting), we gathered a group of people to discuss ideas for using Gaia DR2 data in the service of exoplanet science. We were focusing on easy ideas that could be executed quickly after the data release. These ideas fell into some broad categories. One is to use the astrometry and photometry to get stellar radii and thereby get better estimates of planet radii for the Kepler planets. Another is to use Gaia-based stellar age estimates to compare planetary systems around stars of different ages. Or compare ages for planetary systems of different architectures (as they say). One of my favorite age estimates is the (square of the) vertical action in the Milky-Way disk! Along the same lines: Test theories for pumping or damping of eccentricities with stellar ages. Because of recent work around Flatiron, there was substantial talk of whether Gaia could detect signatures of stars that have recently accreted their planets. There might be different signatures on different time-scales.

Late in the day, I worked some more on my #hackAAS project to look at the dimensionality of stars in element-abundance space. I think (but am not sure) that the best way to think about this problem for the purposes of chemical-tagging applications is in terms of how well we can predict unmeasured abundances. Because this gets at the value or trade-offs between measuring more elements or measuring the elements you already have but better. I need to write first and data-analyze second, but the data are so fun to play with!


statistics arguments

Today Taisiya Kopytova (ASU) showed up to write up our results on element abundances and binarity in the APOGEE survey. My job is to write up some philosophy of the project. The methodology is extremely clean because we have structured it like a properly executed drug trial. I'm excited to propagate this methodology, that is only really possible in big data sets.

Alex Malz (NYU) and I started our discussion of how to finish his project on combining noisy photometric redshifts to obtain the full redshift distribution. It is a hard paper to write because it has a lot of non-trivial conceptual aspects to it. In particular, there is an argument in the literature that you can just stack the posterior PDFs that is wrong (it's provably correct and yet wrong!), and we have to rebut that without getting too much into the weeds.


#hackAAS number 6

Today was the 6th annual AAS Hack Day, at #AAS231 in Washington DC. (I know it was 6th because of this post.) It was an absolutely great day, organized by Kelle Cruz (CUNY), Meg Schwamb (Gemini), and Jim Davenport (UW & WWU), and sponsored by Northrup Grumman and LSST. The Hack Day has become an integral part of the AAS winter meetings, and it is now a sustainable activity that is easy to organize and sponsor.

My hack for the day was to work on dimensionality and structure in the element-abundance data for Solar Twins (we need a name for this data set) created by Megan Bedell (Flatiron). I was reminded how good it is to bring data sets to the Hack Day! Several others took the data to play with, and Martin Henze (SDSU) went to town, visualizing (in R) the covariances (empirical correlations) among the elements. Some of these correlations are positive, some are negative, and some are tiny. Indeed, his analysis sort-of chunks the elements up into blocks that are related in their formation! This is very promising for my long-term goals of obtaining empirical nucleosynthetic yields.

What I did for Hack Day was to visualize the data in the natural (element-abundance) coordinates, and then again in PCA coordinates, where many of the variances really vanish, so the data really are low dimensional (dammit; my loyal reader knows that I don't want this to be true). And then I also visualized in random orthonormal coordinates (which was fun); this shows that the low-variance PCA space is indeed a very rare or hard-to-find subspace in the full 33-dimensional element space. I also visualized some rotations in the space, which forced me to do some 33-dimensional geometry, which is a bit challenging in a room of enthusiastic hackers!

But so much happened at the hack day. There was a project (led by aforementioned Bedell) to make interactive web-based plots of the exoplanet systems, to visualize multiplicity, insolation, and stellar properties. There was a project to find the “Kevin Bacon of astronomy” which was obviously flawed, since it didn't identify yours truly. But it did make a huge network graph of astronomers who use ORCID. Get using that, people! Erik Tollerud did more work on his hack-of-hacks to build great tools for collecting and recording hacks, but he also was working on a white paper for NASA about software licensing. I gave him some content and co-signed it. MIT for the win. Foreman-Mackey led a set of hacks in which astronomers learned how to use TensorFlow, which uses NVIDIA GPUs for insane speed-up in linear-algebra operations. Usually people use TensorFlow for machine learning, but it is a full linear-algebra library, with auto-differentiation baked in.

The AAS-WWT people were in the house, and Jonathan Fay, as per usual at Hack Days (what a hacker!), pushed a substantial change to the software, to make it understand and visualize velocity maps. Another group (including Schwamb, mentioned above) visualized sky polygons in WWT, and used a citizen-science K2 discovery as its test case for visualizing a telescope focal-plane footprint. There were nice hacks with APIs, with people learning to use the NASA Astrophysics Data System API and Virtual Observatory APIs, and getting different APIs to talk together. One hack was to visualize Julia Sets using Julia! It took the room a few minutes to get the joke, but the visualization was great, and very few lines of code in the end. And there were at least two sewing hacks.

None of this does justice: It was a packed room, about 1/3 of the participants completely new to Hack Days, and great atmosphere and energy. I love my job!


nonlinear dimensionality reduction

Today Brice Ménard (JHU) showed me a new dimensionality-reduction method by him and Dalya Baron. He claims it has no free parameters and good performance. But no paper yet!


streams and dust

At Stars Group meeting, Lauren Anderson (Flatiron) and Denis Erkal (Surrey) both spoke about stellar streams. Anderson spoke about finding them with a new search technique that looks at proper motions for stars found in great-circle segments; this is being prepared for Gaia DR2. Erkal spoke about constraining the Milky-Way potential using only configurational aspects of streams: If a small stream segment locally don't contain the Galactic Center, there must be asphericity in the gravitational potential.

Late in the day, Yi-Kuan Chiang (JHU) showed me absolutely beautiful results cross-correlating various Milky-Way dust maps with high-redshift objects. There ought to be no correlations, at least in the low-extinction regions. But there are correlations, and it is because the dust maps are all contaminated by high-redshift dust in the extragalactic objects themselves (or objects correlated therewith). He can conclude nice things about different dust-map techniques. We discussed (inconclusively) whether his work could be turned around and used to improve Milky-Way map-making.

loose talk of #aliens

[Warning: This post is not, strictly, a research post. It is a response to events in the astronomical community in the recent past.] Word on the street (I can't find out, because it is not open) is that some argument broke out on the Facebook(tm) astronomy group about loose discussion on the internets about #aliens, and things like the Boyajian star or the 'Oumuamua asteroid. Since I am partially responsible for this loose talk, here is my position:

First, I want to separate informal discussion (like on twitter or blogs) from formal discussion in scientific papers (like what might be submitted to arXiv) from press releases. These are three different things, and I think we need to treat them differently. Second, I am going to assert that it is reasonable and normal for astronomers to discuss in scientific papers (sometimes) the possibility that there is alien life or alien technology with visible impact on observations. Third, I am going to presume that the non-expert public deserves our complete respect and cooperation. If you disagree with any of these things, my argument might not appeal to you.

On the second assumption (aliens are worthy of discussion), you can ask yourself: Was it a reasonable use of telescope time to look at 'Oumuamua in the radio, to search for technological radio transmissions? If you think that this was a reasonable thing to do with our resources, then you agree with the second assumption. Similarly if you think SETI is worth doing. If you don't think these uses of telescopes are reasonable—and it is understandable and justifiable not to—then you might think all talk of aliens is illegitimate. Fine. But I think that most of us think that it is legitimate to study SETI and related matters. I certainly do.

Now if we accept the second assumption, and if we accept the third assumption (and I really don't have any time for you if you don't accept the third assumption), then I think it is legitimate (and perhaps even ethically required) that we have our discussions about aliens out in the open, visible to the public! The argument (that we shouldn't be talking about such things) appears to be: “Some people (and in particular some news outlets) go ape-shit when there is talk of aliens, so we all need to stop talking about aliens!” But now let's move this into another context: Imagine someone in a role in which they serve the public and are partially responsible for X saying: “Some people (and in particular some news outlets) go ape-shit when there is talk of X, so we all need to stop talking about X!” Obviously that would be a completely unethical position for any public servant to take. And it wouldn't just be fodder for conspiracy theorists, it would actually be evidence of a conspiracy.

Imagine we, as a community, decided to only discuss alien technology in private, and never in public. Would that help or hurt with the wild speculation or ape-shit reactions? In the long run, I think it would hurt us, and hurt the public, and be unethical to boot. Informal discussion of all matters of importance to astronomers are legitimately held in the open. We are public servants, ultimately.

Now, I have two caveats to this. The first is that it is possible for papers and press releases and news articles to be irresponsible about their discussion of aliens. For example, the reportage claiming (example here)—and it may originate in the paper itself—that the reddening observed in the Boyajian Star rules out alien megastructures was debatably irresponsible in two ways. For one, it implied that the megastructure hypothesis was a leading hypothesis, which it was not, and for two, it implied that the megastructure hypothesis was specific enough to be ruled out by reddening, which it wasn't. Indeed, the chatter on Twitter(tm) led to questions about whether aliens could ever be ruled out by observations, and that is an interesting question, which relates to the second assumption (aliens are worthy of discussion) given above. Either way, the paper and resulting press implied that the observational result constrained aliens, which it did not; the posterior probability of aliens (extremely low to begin with) is almost completely unchanged by the observations in that paper. To imply otherwise is to imply that alien technology is a mature scientific hypothesis, which it isn't.

Note, in the above paragraph, that I hold papers and press releases to a higher standard than loose, informal discussion! That is my first assumption, above. You might disagree with it, but note that it would be essentially completely chilling to all informal, open discussion of science if we required refereed-publication-quality backing for anything we say, anywhere. It would effectively re-create the conspiracy that I reject above.

I don't mean to be too critical here, the Boyajian-star paper was overall extremely responsible and careful and sensible. As are many other papers about planet results, even ones that end up getting illustrated with an artist's impression of a rocky planet with ocean shores and/or raging surf. If I have a complaint about exoplanet science as a community (and I count myself a member of this community; I am not casting blame elsewhere), it is about the paper-to-press interface, where artist's conceptions and small signals are amplified into luscious and misleading story-time by perfectly sensible reporters. We (as a community, and as a set of funded projects) are complicit in this.

The second caveat to what I have written above is that I (for one) and many others talk on Twitter(tm) with tongue in cheek and with sarcasm, irony, and exaggeration. It takes knowledge of the medium, of scientists, and of the individuals involved to decode it properly. When I tweeted that it was “likely” that 'Oumuamua was an alien spaceship, I was obviously (to me) exaggerating, for the purposes of having a fun and interesting discussion. And indeed, the asteroid looks different in color, shape, and spin rate (and maybe therefore composition and tensile strength) from other asteroids in our own Solar System. But it might have been irresponsible to use my exaggeration and humor when it comes to aliens, because aliens do set off some people, especially those who might not know the conventions of scientists and twitter. I take that criticism, and I'll try to be more careful.

One last point: The underlying idea of those who say we should keep alien discussion behind closed doors (or cut it off completely) is at least partly that the public can't handle it. I find that attitude disturbing and wrong: In my experience, ordinary people are very wise readers of the news, with good sense and responsibility, and they are just as good at reading arguments on Twitter(tm). The fact that there are some exceptions—or that the Daily Mail is an irresponsible news outlet—does not change the truth of my third assumption (people deserve our respect). We should just ignore and deprecate irresponsible news, and continue to have our discussions out in the open!

In the long run, astronomy will benefit from open-ness, honesty, and carefully circumscribed reporting of goals and results. We won't benefit from hiding our legitimate scientific discussions from the public for fear that they will be mis-interpreted.


stream search

Lauren Anderson (Flatiron) and Vasily Belokurov (Cambridge) have been developing a (relatively) model-free search for compact structures in phase-space, to be used on Gaia DR2 in April. Anderson showed me the current results, which highlight many possible streams, in pre-Gaia test data from SDSS (created by Sergey Koposov at CMU). She requires that the phase-space over-density be consistent with a section of a great circle, at a (somewhat) well-defined Galactocentric distance (because: reflex proper-motion from the Sun's velocity), moving in a direction along the circle. She finds lots of putative streams, but many of them seem to highlight edges and issues with the spatial selection function, so there are still issues to work out. The nice thing is that if we can get something working here, it will crush on the forthcoming Gaia data, which will have a much simpler selection function.


hand-written optimizers

Megan Bedell (Flatiron) and I are working on the optimizers underlying our stellar radial-velocity determination pipeline, code-named wobblé. There were serious bugs! But by the end of the day it looked like everything was optimizing properly.
Note to self: Don't hand-write an optimizer unless it is absolutely necessary!


M-dwarf spectral models; time-variable spectra

I spoke to Jessica Birky (UCSD) today about her #AAS231 poster on using The Cannon to label M-dwarf spectra in the APOGEE spectra. She has beautiful results, using spectral types from the Burgasser group, and using physical labels (temperatures and compositions) from Andrew Mann (Columbia). We discussed things to emphasize on the poster and things to emphasize in discussions with people who come by. For my loyal reader: It will be up at #AAS231 poster session 349 on Thursday January 11.

I also spent some time working on a set of issues around measuring precise radial velocities for stars in the presence of time-variable spectral features in both the star and the tellurics. I worked out derivatives for the spectral model when the telluric absorption is permitted to come from a low-dimensional subspace of spectrum space. I then turned my attention to the ill-posed problem of determining precise radial velocities when the star changes its spectral shape (or line strengths or line positions, etc). In the case that the stellar spectrum changes completely randomly, and independently or separably from the stellar velocity, I believe (oddly) that the problem is going to be easy. The problem is that it won't change completely separably: There will be stellar surface variations that co-vary with spectral changes. This is the reality, and the hardest case, I think.


all papers are about four things

It is pathetic how little work I got done today, given that the International Date Line made today 45 hours long for me. I did get some writing done in my machine-learning opus, and figured out a way to re-structure it to make it more readable. I also thought a bit about the following thing, which came up in discussion with Bedell:

Scientific papers in the astronomy literature are always, and unavoidably, about four things: They are about the Universe or astrophysical phenomena. They are about the data we have on those phenomena. They are about the astronomical literature itself. And they are about the authors themselves. You can't avoid talking about yourself when you write a paper because: subjectivity. All data analyses, frequentist or Bayesian, involve subjective decision-making. You can't avoid talking about the literature, because: relevance. Besides, astronomy is the astronomical literature, in my opinion. And you can't avoid talking about the data, because: accuracy. That is, you must be responding to data in some way or other, even if you are a theorist. And you can't avoid the universe, because otherwise it isn't astronomy!


linear regression and the kernel trick

I spent the day in an undisclosed location, working on linear regression. That is, I am working on the notation and description for linear regression for my opus on machine learning in astronomy. Mid-way through getting it all together, I started to lose faith that I even know what linear regression is—or that what machine learners call linear regression is what I call linear regression! But in my undisclosed location, I don't have a copy of Bishop!

I do have Wikipedia, however, and I spent some time there, reading different descriptions and learning new applications of the kernel trick. I think of it as some kind of “lifting” of the problem to a (far) larger space, but it can also seen as a redefinition of “proximity” or “similarity” in the data space. That makes sense, because (at base) the kernel trick is a redefinition of the dot product. Stuff to think about, and relevant to many machine-learning methods. In particular, when you apply it to linear regression, you get (more or less) the Gaussian Process.


machine learning for astronomers

I spent a bit of vacation time writing in long-term writing projects. The one I found myself wanting to work on is a long-term project called (for now) “Machine learning for astronomers”, in which I go over the basics, and give contextualized (and unsolicited) advice for using machine-learning methods in astrophysics. One of my principal goals is to criticize many uses that fall into the estimator category, and promote methods that can be built into larger, probabilistic inferences. This deprecates most uses of deep learning, and encourages Gaussian Processes. Interestingly, generative adversarial networks (the new rage) are good in this dichotomy, because they are generators that transform probability densities. But I am starting small, working through the detailed mathematics of five methods which I think are so beautiful and simple, everyone should know them.