Citizen statistician is very pleased to announce that one of its own, Andy Zieffler, is this year’s recipient of the American Statistical Association’s Waller Distinguished Teaching Career Award. Congrats, Andy!

# Category Archives: Events

# Community Colleges and the ASA

Rob will be be participating in this event, organized by Nicholas Horton:

CONNECTION WITH COMMUNITY COLLEGES: second in the guidelines for undergraduate statistics programs webinar series

The American Statistical Association endorses the value of undergraduate programs in statistical science, both for statistical science majors and for students in other majors seeking a minor or concentration. Guidelines for such programs were promulgated in 2000, and a new workgroup is working to update them.

To help gather input and identify issues and areas for discussion, the workgroup has organized a series of webinars to focus on different issues.

**Connection with Community Colleges**

Monday, October 21st, 6:00-6:45pm Eastern Time

Description: Community colleges serve a key role in the US higher education system, accounting for approximately 40% of all enrollments. In this webinar, representatives from community colleges and universities with many community college transfers will discuss the interface between the systems and ways to prepare students for undergraduate degrees and minors in statistics.

The webinar is free to attend, and a recording will be made available after the event. To sign up, please email Rebecca Nichols (rebecca@amstat.org).

More information about the existing curriculum guidelines as well as a survey can be found at:

# JSM 2013 – Days 4 and 4.5

I started off my Wednesday with the “The New Face of Statistics Education (#480)” session. Erin Blackenship from UNL talked about their second course in statistics, a math/stat course where students don’t just learn how to calculate sufficient statistics and unbiased estimators but also learn what the values they’re calculating mean in context of the data. The goal of the course is to bring together the kind of reasoning emphasized in intro stat courses with the mathematical rigor of a traditional math/stat course. Blackenship mentioned that almost 90% of the students taking the class are actuarial science students who need to pass the P exam (the first actuarial exam) therefore the probability theory must be a major component of the course. However UNL has been bridging the gap between these demands and the GAISE guidelines by introducing technology to the course (simulating empirical sampling distributions, checking distributional assumptions, numerical approximation) as well as using writing assessments to improve and evaluate student learning. For example, students are asked to explain in their own words the difference between a sufficient statistic and minimal sufficient statistic, and answers that put things in context instead of regurgitating differences are graded highly. This approach not only allows students who struggle with math to demonstrate understanding, but it also reveals shallow understanding of students who might be testing well in terms of the math by simply going through the mechanics.

In my intro stat class I used to ask similar questions on exams, but have been doing so less and less lately in the interest of time spent on grading (they can be tedious to grade). However lately I’ve been trying to incorporate more activities into the class, and I’m thinking such exercises might be quite appropriate as class activities where students work in teams to perfect their answers and perhaps even teams then grading each others’ answers.

Anyway, back to the session… Another talk in the session given by Chris Malone from Winona State was about modernizing the undergraduate curriculum. Chris made the point that we need much more than just cosmetic changes as he believes the current undergraduate curriculum is disconnected from what graduates are doing when they get their first job. His claim was that the current curriculum is designed for the student who is going on to graduate school in statistics, but that that’s only about a fifth of the students in undergraduate majors. (As an aside, I would have guessed the ratio to be even lower.) He advocated for more computing in the undergrad curriculum, a common thread among many of the education talks at JSM this year, and described a few new programs at Winona and other universities on data science. Another common thread was this discussion of “data science” vs. “statistics”, but I’m not going to go there – at least not in this post. (If you’re interested in this discussion, this Simply Statistician post initiated a good conversation on the topic in the comments section.) I started making a list of Data Science programs I found while searching online but this post seems to have a pretty exhaustive list (original post dates back to 2012 but it seems to be updated regularly).

Other notes from the day:

- R visreg package looks pretty cool, though perhaps not necessarily very useful for an intro stat course where we don’t cover interactions, non-linear regression, etc.

- There is another DataFest like competition going on in the Midwest: MUDAC – maybe we should do a contributed session at JSM next year where organizers share experiences with each other and the audience to solicit more interest in their events or inspire others.

On Thursday I only attended one session: “**Teaching the Fundamentals (#699)**” (the very last session, mine). You can find my slides for my talk on using R Markdown to teach data analysis in R as well as to instill the importance of reproducible research early on here.

One of the other speakers in my session was Robert Jernigan, who I recognize from this video. He talked about how students confuse “diversity” and “variability” and hence have a difficult time understanding why a dataset like [60,60,60,10,10,10] has a higher standard deviation than a dataset like [10,20,30,40,50,60]. He also mentioned his blog statpics.com, which seems to have some interesting examples of images like the ones in his video on distributions.

John Walker from Cal Poly San Luis Obispo discussed his experiment on how well students can recognize normal and non-normal distributions using normal probability plots — a standard approach for checking conditions for many statistical methods. He showed that faculty do significantly better than students, which I suppose means that you do get better at this with more exposure. However the results aren’t final, and he is considering some changes to his design. I’m eager to see the final results of his experiment, especially if they come with some evidence/suggestions for what the best method to teach this skill is.

# JSM 2013 – Day 3

Tuesday was a slightly shorter day for me in terms of talks as I had a couple meetings to attend. The first talk I attended was my colleague Kari Lock Morgan’s talk titled “Teaching PhD Students How to Teach” (in the “Teaching Outside the Box, Ever So Slightly (# 358)” session). The talk was about a class on teaching that she took as a grad student and now teaches at Duke. She actually started off by saying that she thought the title of her talk was misleading, as the talk wasn’t about teaching PhD students a particular way to teach, but instead about getting these students to think about teaching, which, especially in research universities, can take a backseat to research. This course features role playing office hours, video-taped teaching sessions which students then watch and critique themselves and each other, as well as writing and revising teaching statements. If you’re interested in creating a similar course, you can find her materials on her course webpage.

In the afternoon I attended part of the “The ‘Third’ Course in Applied Statistics for Undergraduates (#414)” session. The first talk titled “Statistics Without the Normal Distribution” by Monnie McGee started off by listing three “lies” and corresponding “truths”:

- Lie: T-intervals are appropriate for n>30.
- Truth: It’s time to retire the n>30 rule. (She referenced this paper by Tim Hesterberg.)

- Lie: Use the t-distribution for small data sets.
- Truth: Permutation distributions give exact p-values for small data sets.

- Lie: If a linear regression doesn’t work, try a transformation.
- Truth: The world is nonlinear and multivariate and dynamic. (I don’t think “try a transformation” should be considered a lie, perhaps a “lie” would be “If a linear regression doesn’t work, a transformation will always work.”)

McGee talked about how they’ve reorganized the curriculum at Southern Methodist University so that statistics students take a class on non-parametrics before their sampling course. This class covers rank and EDF-based procedures such as the Wilcoxon, signed rank, and Mann-Whitney tests as well as resampling methods which are especially useful for estimation of numerous features of a distribution, like the median, independently of the population distribution. The course uses the text by Higgins (Introduction to Modern Nonparametric Statistics) as well as a series of supplements (which I didn’t take notes on, but I’m sure she’d be happy to share the list with you if you’re interested). However she also mentioned that she is looking for an alternative textbook for the course. Pedagogically, the class uses just in time teaching methods — students read the material and complete warm up exercises before class each week, and class time is tailored to concepts that students appear to be struggling with based on their performance on the warm up exercises.

The second talk in the session titled “Nonlinear, Non-Normal, Non-Independent?” was given by Alison Gibbs. Gibbs also described a course that focuses on models for situations when classical regression assumptions aren’t met. She gave examples from a case study on HPV vaccinations that she uses in this class (I believe the data come from this paper). She emphasized the importance of introducing datasets that are interesting, controversial, authentic, and that lend themselves to asking compelling questions. She also mentioned that she doesn’t use a textbook for this class, and finds this liberating. While I can see how not being tied to a textbook would be liberating, I can’t help but think some students might find it difficult to not have a reference — especially those who are struggling in the class. However I presume this issue can be addressed by providing the students with lecture notes and other resources in a *very* organized fashion. I have to admit that I was hoping that I would hear Gibbs talk about her MOOC at this conference as I am gearing up to teach a similar MOOC next year. Perhaps I should track her down and pick her brain a bit…

At this point I ducked out of this session to see my husband Colin Rundel’s talk in the “Statistical Computing: Software and Graphics (#430)” session. His talk was on a new R package that he is working on (RcppGP) to improve the performance of Gaussian process models using GPU computing. He started with a quote: “If the computing complexity is linear, you’re OK; if quadratic, pray; if cubic, give up.” Looks like he and other people working in this area are not willing to give up quite yet. If you’re interested in his code and slides, you can find them at his GitHub page.

The sessions on my agenda for tomorrow are:

- 8:30am: The New Face of Statistics Education (#480)

- 10:30am: Topics in Statistical Graphics (#543) - specifically the talk on the visreg package as I’m looking for good (and simple) visualizations for simple and multiple regression that I can use in my introductory statistics course.
- 2pm: Assessing the Statistical Understanding and Reasoning in K-12: The LOCUS Project (#587)
- 4pm: COPSS Awards and Fisher Lecture: From Fisher to Big Data: Continuities and Discontinuities (#612)

# JSM 2013 – Day 2

My Monday at JSM started with the “The Profession of Statistics and Its Impact on the Media (#102)” session. The first speaker in the session, Mark Hansen, was a professor of mine at UCLA, so it was nice to see a familiar face (or more like hear a familiar voice – the room was so jam packed that I couldn’t really “see” him) and catch up on what he has been working on at his new position at Columbia University as a Professor of Journalism and the Director of David and Helen Gurley Brown Institute for Media Innovation. The main theme of the talk was the interaction between journalists and statisticians — he discussed how journalism can provide much needed perspective, language, and practices necessary to describe the forces that data exert in our worlds, to help even statisticians gain fresh perspective on their practice. He pointed out a difference between how journalists and statisticians work with data: journalists work with data to tell a story in the context of a dataset, while statisticians tend to tell a story of the dataset. Hansen also discussed Columbia’s new two-year dual degree Master’s in journalism and computer science. The Brown Institute also awards seed funding to students for developing media technologies that could transform how news content is produced, delivered and consumed. I’ve listed a few of the projects that Hansen discussed below, and a detailed post on these grants can be found here.

- Dispatch: a mobile application that provides secure, authenticated, anonymous instant publishing.
- Personalized Television News: a project that seeks to develop and demonstrate a platform for personalized television news to replace the traditional one-broadcast-fits-all model.

- CityBeat: a project that looks for newsworthy events in the patterns of real-time, geotagged social media feeds.
- The Declassification Engine: An engine that uses machine learning to declassify documents.
- Bushwig: Telling the story of a drag renaissance taking place in Bushwick, Brooklyn, that is enlisting and extending social media platforms for the “identity curation” that happens in the drag community. I had no idea that Facebook does not allow, or at least takes down when found out, two profiles for the same person, which, as you can imagine, can be an issue for people who live their lives in two identities.

Hansen also discussed a recent project where he collaborated with the NYTimes’ R&D lab, working on projects such as Project Cascade, which is a tool that constructs “a detailed picture of how information propagates through the social media space”, like Twitter.

The next talk in the session by Don Berry discussed fundamental issues in statistics that are difficult to convey to journalists, and hence the rest of the public, such as Simpson’s paradox, results that are “too good to be true” (e.g. dogs sniffing cancer), regression to the mean, multiple comparisons, etc. He also discussed at length prosecutor’s fallacy, within the context of the case of nurse Lucia de Berk who was convicted in was convicted in 2004 of a number of murders and attempted murders of patients in her care, but then was freed in 2010. I don’t discuss prosecutor’s fallacy in my introductory statistics class, but I’m thinking that I should… Berry recommended this NYTimes article on the topic, as well as this TED talk on prosecutor’s fallacy in general). Berry, who is often quoted in newspaper articles as an expert, also discussed what statisticians (and other scientists) should and should not do when interacting with journalists. Some of the key points were:

- Simplify, short of lying
- Be pithy
- Avoid questions that you don’t want to answer – He mentioned that he avoids questions like “What are the economic implications?”
- Use going off the record sparingly
- Prefer email over telephone – so that you can edit your own words
- Don’t diss anyone

The last one seems obvious, but see this this Washington Post article on the 2009 breast cancer screening frequency controversy. In the article, a radiology professor from Harvard is quoted saying “Tens of thousands of lives are being saved by mammography screening, and these idiots want to do away with it”. Wow!

The next speaker was Howard Wainer (whose article titled ”The Most Dangerous Equation: Ignorance of how sample size affects statistical variation has created havoc for nearly a millennium” is a good read, by the way). I am excited to take a peek at his recently published book Medical Illuminations at the Expo later today.

The last speaker in the session was Alan Schwarz, the Pulitzer-prize nominated reporter at the NYTimes who did an expose on current and retired football players suffering from post-concussion syndrome and early-onset dementia, more specifically Chronic Traumatic Encephalopathy. A journalist talking about using data and statistics to uncover a story was a nice complement to the earlier talks by statisticians talking about working with journalists.

In the afternoon I attended the “Toward Big Data in Teaching Statistics (#210)” session. Nicholas Chamandy from Google talked about how big data requires novel solutions to old problems and gave examples from some of the algorithms Google uses to solve problems in predictive modeling.

Randall Pruim’s talk focused on the efforts of the Computation and Visualization Consortium that has started working on identifying key skills that students need to work with big data and ways to teach them. It was quite eye opening to hear about a survey he conducted asking faculty members from science departments such as physics and chemistry what kind/size of data their students work with – turns out for many the answer is no data at all! He also gave an overview of efforts at Macalester, Smith, and Calvin Colleges for introducing big data skills into their curriculum. I will be looking into the syllabus for the class being taught at Macalester by Danny Kaplan, as I’m also currently brainstorming how best to teach core computational skills to our students.

Nick Horton also discussed his vision for accomplishing this, which is to start in the first course, to build on it in the second course, to provide more opportunities for students to apply their knowledge in practice (internships, collaborative research, teaching assistants), and to introduce new courses focused on data science into the curriculum. He also discussed exposing students to reproducible research using RStudio and R Markdown. I’ve previously written a blog post about this, and will be talking about it on Thursday at my own talk as well. It was nice to see a similar approach being used by others in the statistics education field. What especially resonated with me was Nick’s comment on how using R Markdown facilitates appropriate and correct statistical workflow for students.

The last talk of the day I attended was Nate Silver’s President’s Invited Address, along with just about everyone else attending JSM. The turnout was great, and his talk was highly enjoyable, as expected. Gregory Matthews (Stats in the Wild) already posted a list of his talking points, so instead of listing them here again, I’ll just link to that post. The Q&A was just as interesting as the talk itself, below are a few notes I jotted down:

- Q: What can statisticians learn from journalists?
- A: Clarity of expression – results are only useful when you can explain them.

- Q: How can ASA and statisticians do more on advocacy?
- A: Blog! Researchers should do their own communication.

- Q: Any career advice for young statisticians?
- A: Do something practical and applied first, theory is easier to learn as needed.

- Q: Favorite journalist/writer?
- A: Bill James

- Q: Data scientist vs. statistician?
- A: Call yourself whatever you want, just do good work. (One of the better answers I’ve heard on this topic. Though his earlier answer “data science is just a sexed up term for statistics” seemed to resonate well with some in the room and not so much with others.)

- Q: What is the future of sports statistics?
- A: More data being collected on soccer, so there is more to be done there. (This means that finally there may be sports statistics that I actually care about and can get excited by!)

After the talks I stopped by the UCLA mixer, it was nice to see some old faces. And I finished up the evening at the Duke dinner, with great company and lots of wine…

Now on to Day 3…

# JSM 2013 – Day 1

Bonjour de Montréal!

I’m at JSM 2013, and thought it might be nice to give a brief summary of highlights of each day. Given the size of the event, any session that I attend means I’m missing at least ten others. So this is in no way an exhaustive overview of the day at the conference, more tidbits from *my* day here. I’ll make a public commitment to post daily throughout the conference, hoping that the guilt of not living up to my promise helps me not lose steam after a couple days.

The first session I attended today had a not so exciting title — “Various Topics in Statistics Education (#43)” — but turned out to be quite the opposite. The first three talks of the session were about the case of Diederik Stapel – a former professor of social psychology in the Netherlands who was suspended from Tilburg University for research fraud. Stapel published widely publicized studies, some of which included results that purport to show that a trash-filled environment tended to bring out racist tendencies in individuals or that eating meat made people selfish and less social. As the speakers at the session (Ruud Koning, Marijtje van Duijn, and Wendy Post from the University of Groningen, and Don van Ravenzwaaij from the University of New South Wales) put it today, the data and the results were “too good to be true”.

First, Koning gave an overview of the case – unfortunately I walked in a little late. If you’re not familiar with it, I would recommend this NYTimes article as well as this paper by Pieter Drenth.

Next, van Duijn discussed best practices for reviewers so that fraud can be caught early on. For example, some indicators of mistakes in Stapel’s papers were impossible means and effect sizes (compared to previous literature), impossible combinations of sample size and degrees of freedom, and incorrect p-values. These could, and should, have been caught by reviewers but this is easier said than done. van Dujin and van Ravenzwaaij suggest that journals should encourage sharing data and reproducibility (a view shared by many in the statistics community). However the responsibility of ensuring thorough reviews should also be shared by universities, science foundations, and policy makers. For example, an interesting suggestion was universities rewarding good peer reviews, as well as good data collection, archiving, and sharing.

The last talk in the series given by Post focused on what to do in education to prevent fraud. Two points that resonated with me were the need for teaching data management, as early as possible in the curriculum, and focusing on descriptive statistics before p-values. Post also advocates for putting emphasis on teaching philosophy of science.

Not only was this discussion very informative and interesting to listen to, it also provided me with a good case study to incorporate into my Statistical Consulting course which has a research ethics component. In the past year we’ve discussed the Potti case, so this will be a nice addition from a different field (and a different university!).

The other session that I attended today was the “Introductory Overview Lecture: Celebrating the History of Statistics (#47)” by Xiao-Li Meng, Alan Agresti, and Stephen Stigler. If you are interested in history of statistics departments, Agresti and Meng’s book *Strength in Numbers: The Rising of Academic Statistics Departments in the US* sounds like a promising read. The session wrapped up with Stigler’s history of statistics review, titled “How Statistics Saved the Human Race”. He was a delight to listen to as usual. I hope that in the future such sessions are recorded and posted online for all to see, as they should be of interest to a wide audience of statisticians and non-statisticians alike. I don’t think the ASA does this yet, but correct me if I’m wrong.

Three other sessions that I would like to have attended today were

- “Teaching Ethics in Statistics and Biostatistics: What Works, What Doesn’t Work, and Lessons Learned (#55)”,
- “The Interplay Between Consulting and Teaching (#68)”, and
- “Teaching Online on a Budget (#75)”.

If you’ve been to any of these, and have notes to share, please comment below!

On a separate note, unrelated to JSM –

- If you’re here in Montréal, and especially if you live in a city without good bagels (Durham, I love you, but you don’t deliver on this account), I strongly recommend a trip up to Fairmount Bagel. They’re open 24 hours, and the bagels are great, but the matzoh bread is to die for. Also, apparently in Quebec “everything” bagels are called “all dressed”.
- It turns out that not everything is good with maple syrup. I strongly advise against trying the Lay’s Maple Moose chips. Trust me on this one.

# DataFest 2013

DataFest is growing larger and larger. This year, we hosted an event at Duke (Mine organized this) with teams from NCSU and UNC, and at UCLA (Rob organized) with teams from Pomona College, Cal State Long Beach, University of Southern California, and UC Riverside. We are very grateful to Vaclav Petricek at eHarmony for providing us with the data, which consisted of roughly one million “user-candidate” pairs, and a couple of hundred variables including “words friends would use to describe you”, ideal characteristics in a partner, the importance of those characteristics, and the all-important ‘did she email him’ and ‘did he email her’ variables.

The students had a great time, and worked hard for 48 hours to prepare short presentations for the judges. This is the third year we’ve done this, and I’m growing impressed with the growing technical skills of the students. (Which makes our life a lot easier, as far as providing help goes.) Or maybe it’s just that I’ve been lucky enough to get more and more “VIP Consultants” (statisticians from off-campus) and talented and dedicated grad students to help out, so that I can be comfortably oblivious to the technical struggles. Or all of the above.

One thing I noticed that will definitely require some adjustment to our curriculum: Our students had a hard time generating interesting questions from these data. Part of the challenge is to look at a large, rich dataset and think “What can I show the world that the world would like to know?” Too many students went directly to model-fitting, without making visuals or engaging in the content of the materials (a surprise, since we thought they would find this material much more easily-engageable than last year’s micro-lending transaction data), or strategizing around some Big Questions. They managed to pull it off in the end, most of them, but would have done better to brainstorm some good questions to follow, and would have done much better to start with the visuals.

One of the fun parts of DataFest is the presentations. Students have only 5 minutes and 2 slides to convince the judges of their worthiness. At UCLA, because we were concerned about having too many teams for the judges to endure, we had two rounds. First, a “speed dating” round in which participants had only 60 seconds and one slide. We surprised them by announcing, at the start, that to move onto the next round, they would have to merge their team with one other team, and so these 60-second presentations should be viewed as pitches to potential partners. We had hoped that teams would match on similar-themes or something, and this did happen; but many matches were between teams of friends. The “super teams” were then allowed to make a 5-minute presentation, and awards were given to these large teams. The judges gave two awards for Best Insight (one to a super-team from Pomona College and another to a super-team from UCLA) and a Best Visualization (to the super-team from USC). We did have two inter-collegiate super-teams (UCLA/Cal State Long Beach and UCLA/UCR) make it to the final round.

If you want to host your own DataFest, drop a line to Mine or me and we can give you lots of advice. And if you sit on a large, interesting data set we can use for next year, definitely drop us a line!

# This Day in Statistics

I was looking to find an add-on Google Calendar that included important days in the history of statistics. They have one for seemingly everything under the sun, except this. So I created one and made it public in honor of the International Year of Statistics. I will continually add to it as I find time.

Feel free to add it. As always, it is available in the following formats

Want me to add an important birthday? Add the info into the comments section. Want to be an author on the calendar so you can add all 100 statistician’s that you know I forgot? Send me an email and I will add you on.

# NCTM Essential Understandings

NCTM has finally published books on statistics in its EU series. This is a rather traditional approach to statistics, given the context of this blog. But, since I’m a co-author (along with Roxy Peck and Stephen Miller), why not point you to it?

http://www.nctm.org/catalog/product.aspx?ID=13804

And while the book is not computational in theme, it does address a central issue of this blog: universal statistical knowledge.

A grades 6-9 version is due out any moment. Stay tuned.

# New Flowing Data Book

http://flowingdata.com/2013/01/22/data-points-first-look/

Nathan Yau has a new book coming out, this about working with data. Pre-order now!