Reflections on the Baltimore Riots

Let me preface this by saying I didn’t sleep much last night, so my ideas are all half-baked. I am tempted to blame my insatiable curiosity about the Baltimore riots on the industry I work in, but it may be the other way around. Regardless, I could not look away from Twitter last night despite the horrible images, violent videos, and – worst of all – the oversimplified commentary coming from people around the country.

I knew immediately how I didn’t feel about the riots: I wasn’t excited or vindicated by the people I saw taking the streets with violence, nor did I completely condemn those involved. But until now, I haven’t been able to articulate how I do feel.

It came to me this morning. A friend and I spoke about the tragic events unfolding with all the perspective of one restless night.

“I don’t feel like I can have an opinion because I’m not there,” she told me.

For context, this friend lives in Charleston, South Carolina. You may remember, North Charleston recently made national news headlines when the New York Times released a video showing a white police officer shooting a fleeing, unarmed black man.

I, on the other hand, am in Rochester, New York. Yesterday I was in court covering the trial of a young black man who shot and killed a white police officer during a foot chase.

Maybe these incidents, taken at face value, are not directly related, but they and others are all connected. Those of us in other cities across the United States watched as Baltimore burned. We experienced emotion across the spectrum from heartbreak to frustration to anger, and some of us wonder if it could happen here.

We live in cities where income disparity and racial segregation draw lines criss-crossing through our neighborhoods but we look at places like Ferguson, North Charleston, Baltimore as if they are worlds away.

I don’t know the answer to the deep-rooted, systemic problems that plague cities across this country – how do you end something like poverty? – but I know that burying our heads in the sand and saying NIMBY is only making matters worse.

When the fires are put out and the rioters go home and the media frenzy that has swarmed up around these tragic events finally quiets, we cannot forget how this started, because what happened in Baltimore not only could happen anywhere, but it already is.

UPDATE: Here are some people that say stuff and write things better than I do:

Jelani Cobb’s Baltimore and the State of American Cities

Ta-Nehisi Coates’ Nonviolence as Compliance


Sentiment Analysis

crazy head gearI’ve been following some cool work coming out of University of Rochester. I did a story for WXXI here, about how researchers in the computer science department teamed up with researchers in the medical department to create software that could help track your mood and monitor mental health.

I spoke with Jiebo Luo, an engineer, and Vince Silenzio, a psychiatrist. To most people, the most notable aspect of their research is the selfie-for-your-health-y video feature. Users take a vid of their faces with a front-facing camera and a computer tells them what they’re emoting. The software reads cues like facial ticks, micro-expressions, and other nonverbal cues. It can gauge your heart rate by analyzing your forehead. It can register the dilation of your pupils. And, of course, it can read whether or not you’re smiling. (Victims of BRF beware.)

appThe less-sexy part of this software, which Luo says could potentially be turned into a smartphone app, has nothing to do with videos. The program also analyzes its users social media use. Everything from how fast your scrolling to the kitten picture you reposted is judged. It also analyzes your network, and puts a value on every link that connects that network, from friendships between users, to likes, to posts or comments.

Right now, the program only registers emotions as positive, negative, or neutral. As far as algorithms are concerned, every interaction you have with your social media has a value of +, -, or n.

So let’s break this down:

Let’s say you’re strolling around Facebook town and you see a picture of a kitten that you just have to share on your wall. The computer would analyze the picture. (Fluffy, baby animal? Pinkish hues? Bright light? Value: Positive.) You added some text with the picture: AWWWWW! (Value: Positive) You get a ton of likes and a few smiley emoticons in the comments section (Positive, positive). Odds are, your day is going pretty sweet.

Mental Health and Social Networks

For someone battling mental health issues like depression or bipolar disorder, this information could really help them get a handle on what’s going on inside their heads. It could also help bridge the gap between a patient and a clinician during a tele-health session, which is a story for another post.

A lot of work went into analyzing that kitten picture. Luo’s been working on sentiment analysis for some time. His algorythm, called Sentribute, does exactly as the name suggests. It uses attributes within an image to judge its sentiment. I won’t go into the minutia of the operations (mostly because I tried really hard but I really don’t understand any of them I’m so sorry) but you can read his paper on it here.

But Luo doesn’t stop there. Not only can you use sentiment analysis to determine the mood of an individual, you can use it to monitor the mood of the country!

I’m not being hyperbolic. Stay with me.

Luo took his fancy-pants algorithm to Flickr and went through images of 2008 presidential candidates. He wanted to see if social (multi)media could be used to predict elections. His theory was that people will post pictures of the candidates they like and will vote for, and therefore the candidate with the greatest online showing would be the victor.

In a sense, each time an image or video is uploaded or viewed, it constitutes an implicit vote for (or against) the subject of the image. This vote carries along with it a rich set of associated data including time and (often) location information. By aggregating such votes across millions of Internet users, we reveal the wisdom that is embedded in social multimedia sites for social science applications such as politics, economics, and marketing.

During this process, he had to correct for things like sacasm and attack ads, which is where the sentiment analysis comes in. Not all pictures of Obama are counted as “votes” per-se, but flattering pictures of him (smiling, head up, arms raised, otherwise stereotypically strong, powerful) would be.


The results? Flickr users correctly predicted the outcome of the primary and general elections of 2008. Sometimes, with super accuracy that I feel couldn’t be recreated if you tried.

An interesting finding: Obama won the presidential election on 11/4/2008 by capturing 52.9% of the popular vote [19]. The TUPD score for Obama and McCain on the same day is 642 (Obama) v.s. 571 (McCain) [F], with Obama got 642/(642 + 571) = 52.9% exactly the same as the real voting result.

You may be asking yourself, Would this experiment have been as accurate if it didn’t feature Barack Obama, “The Social Media President?” But Luo’s research also determined that Flickr photos accurately predicted the Iowa caucuses, the New Hampshire Primary, and the Michigan primary wins for Huckabee, McCain, and Romney, respectively.

On this kind of macro-level, Luo thinks this research can lead to a greater understanding of the wisdom of crowds. He also thinks it can be used to understand economics and marketing in a new way. (His paper includes analysis of Apple product sales through Flickr shares.)

But on a smaller scale, Luo’s research can give us all kinds of insights about the pictures we choose to post and share. Maybe your decision to use a particular Instagram filter is a warning sign that your on the edge of a depressive episode and could warn you to get help before it becomes an issue. It could help us understand how we deal with contagious emotions within our personal networks. As Luo told me, the applications are endless!

Sentiment analysis is kind of blowing my mind right now, in part because recognizing emotions is so embedded in my understanding of the human experience. I’ll categorize this under “turning computers into sentient beings.”

Also, I got through this entire post without stating the obvious “A PICTURE IS WORTH A THOUSAND WORDS” joke (until just now).


The Role of Narrative

Zoomorphic Calligraphy

Zoomorphic Calligraphy

It happens when you least expect it. One minute, you’re asking an anti-terrorism expert about global Islamic extremism and Islamophobia in the media. The next, you’re discussing identity and empathy through storytelling, and the importance of individual and cultural narratives. I love my job.

Mark Concordia is a professor of criminal justice at Roberts Wesleyan College.

TL;DR: Internet Use and “ADD”


Where’s my soapbox? Oh, there it is. Ok:

We all have some kind of ADD now. We all need to stop being mean to each other about it. And we all need to find ways to turn it around because it’s a total bummer and I don’t want my generation to turn into a bunch of over-medicated zombie people.

The DSM V gives the behavioral disorder known as Attention Deficit (Hyperactivity) Disorder the following criteria:

  1. Often does not give close attention to details or makes careless mistakes in schoolwork, work, or other activities (e.g., overlooks or misses details, work is inaccurate)
  2. Often has difficulty sustaining attention in tasks or play activities (e.g., has difficulty remaining focused during lectures, conversations, or reading lengthy writings).
  3. Often does not seem to listen when spoken to directly (e.g., mind seems elsewhere, even in the absence of any obvious distraction)
  4. Often does not follow through on instructions and fails to finish schoolwork, chores, or duties in the workplace (e.g., starts tasks but quickly loses focus and is easily sidetracked; fails to finish schoolwork, household chores, or tasks in the workplace)
  5. Often has difficulty organizing tasks and activities (e.g., difficulty managing sequential tasks; difficulty keeping materials and belongings in order; messy, disorganized, work; poor time management; tends to fail to meet deadlines)
  6. Often avoids, dislikes, or is reluctant to engage in tasks that require sustained mental effort (e.g., schoolwork or homework; for older adolescents and adults, preparing reports, completing forms, or reviewing lengthy papers).
  7. Often loses things needed for tasks and activities (e.g., school materials, pencils, books, tools, wallets, keys, paperwork, eyeglasses, or mobile telephones)
  8. Is often easily distracted by extraneous stimuli (for older adolescents and adults, may include unrelated thoughts).
  9. Is often forgetful in daily activities (e.g., chores, running errands; for older adolescents and adults, returning calls, paying bills, keeping appointments)

Of the nine criteria listed above, a patient need display 6 or more for 6 months to be diagnosed with ADD. The New York Times reports a few different statistics on the recent rise of ADD including: 11% of school age children are diagnosed, 1 in 5 high school age boys are diagnosed, etc.

There are a plethora of opinions about why this problem, perceived or real, is on the incline. Some think a reduced stigma in psychotherapy encourages more people to seek help, leading to an influx of diagnoses. Readers of Allen Francis’ Saving Normal might say that drug-pushing Pharmaceutical companies are to blame. It is also not uncommon for those on medication to be accused of faking. And there are a few hippie weirdos out there who point to Food Inc. All of these theories have their own level of credibility (some more than others), but I’m going to ditch them all and make a totally non-credible undergraduate-level hypothesis based on surface-level investigation, Google, personal experience, and one book I read one time. I took Psych 101 in Community College, so I’m going to make a sweeping diagnosis on everyone who will ever read this page. Stay tuned.

Personal Experience

I always thought the people I knew who were diagnosed with Adult ADD were gaming the system; thought they were trying to get one over on their doctors so they could have a leg-up during finals week or write a term paper in one weekend. Once my friends entered the work force, some continued taking ADD meds to set the world on fire, and my opinion shifted. I thought, Why should it be easier for them when it’s so hard for the rest of us? Because, in my own experience, it is hard for the rest of us. I could continue to blame people for exploiting a disorder to function more productively, or I could embrace that maybe we all have it, to varying degrees.


I’m not good with all the science-y stuff, but luckily I don’t have to be. Nicholas Carr wrote a little book called The Shallows that lays it all out and, besides being a little gratuitously repetitious, it’s a good read. Over-simplification time: The book makes the case that increased exposure to the Internet is fundamentally changing the way our brains work. He explains in detail the theories of plasticity of our neurons and brain patterns, and their susceptibility to change throughout our lives due to the habits we form. In the now out-dated print culture, experience with books and other print media created a population of people capable of more prolonged attention (generally speaking). Reading was good exercise for the brain, like little brainy push-ups, keeping it focused in a linear way on a specific task, and making prolonged attention to other tasks (work, homework, life stuff) easier.

Television came in and stirred the pot a little, but the Internet is what screwed it all up for everybody. In the same way reading trained the brain to pay attention, interacting with the Internet trains the brain to follow tangents. Hyperlinks, multiple tabs, and second screen interaction all contribute to this phenomenon in the same way. If I haven’t lost you yet, maybe you’re beginning to see the problem. Because our minds are reshaping to acclimate to the tools we use (computers, smart phones, etc.), we are suffering from attention problems and memory loss. You can’t hyperlink your way through real life if you want to have successful relationships and careers.

In his article for the Huffington Post, Dr. Mark Goulston says, “When your thinking is interrupted by your brain, you’ve got real ADD. When it’s interrupted by the world, you just have trouble saying ‘no.’” I say nay, Dr. Goulston. More specifically, I say since we’ve trained our brains out of the ability of saying “no,” your distinction is imperceptible. Using words like “pseudo-ADD” does not change the experience of people who live in a world that demands their attention while their mind simultaneously demands distraction.

Patronizing Each Other is Getting Us Nowhere

In researching this post I came across a multitude of articles, not unlike Mark Goulston’s, that take a decidedly condescending tone when talking about “ADD fakers” (although they fail to accurately define the difference between them and their non-faking-counterparts). I was interested, though unsurprised, to see the authors of such articles attributed attention problems to shortcomings on the behalf of the spacey person. Lori Day writes, rather facetiously, “I know, I know, it’s all to be blamed on our fast-paced, screen media-saturated, multi-tasking lifestyles. We can’t help it. There’s just too much going on.” Well, excuse me Miss Day. I didn’t mean to bore you.

My guess is, this harsh attitude is the result of two completely separate viewpoints. Theory Number One: People are frustrated with their own attention deficits, but do not use a label like “ADD” to excuse their behavior and look down on those who they think do so. Theory Number Two: Neuroplasticity may account for the ongoing change of our brains, but lots of habits (good and bad) are formed in childhood. These attention problems may be generational, with younger people facing more obstacles to their productivity because they did not develop more focused habits as kids (they were too busy with other stuff) while their older counterparts may be more disciplined and not-so-understanding.

Get over it.

Stop implying that people are faking having ADD or saying things like pseudo-ADD and listen (look?) up. We need to drop this whole label, non-label nonsense and get to the crux of the issue. Unless we all throw down our digital devices in a show of solidarity (hahahaha), this is the new normal. We need to find ways to combat the attention problems that affect our careers and personal lives, instead of being condescending to each other or worse, taking meds for it. This is my one and only disclaimer for this post: I can’t say that meds don’t help anybody, because maybe they do. But then again, I don’t think I know anybody with a form of ADD that wouldn’t benefit from some alternative treatment.

Concentration and extended focus are important for obvious reasons.  The ability to pay attention leads to productivity at school and in the workplace (read: more $$$).  But there are more important things at stake here.  As a society we are moving rapidly toward a highly digitized world, and shorter attention spans are only the tip of the iceberg.  Institutions founded on the values of print culture are suffering, becoming more obsolete and functioning poorly.  It takes a certain kind of mind to contemplate complex problems like a broken political structure, a failing education system, and unsustainable energy consumption across the globe.  We all need to strive for this certain kind of mind, before we lose ourselves to the distractions completely.

I have a point I promise. I read an awesome article on Lifehacker about rebuilding your attention span. After a few clicks I found a bunch more articles that deal with the same issue. We need to combat the impact the internet is having on our brains – making us inattentive, forgetful, anxious, unmotivated, and unproductive – in order to lead happier, healthier lives. We can do this. But forming healthier habits isn’t going to happen over night, and although I think it’s possible it would definitely require a lot of effort, and…

Oh, look, a video.

“Bros” Before Other Webshows

See what I did there?

Hello, strangers! I have come out of hiding to tell you about this awesome youtube video that you have to watch yesterday:

(I heard about this almost-web series from some staff at school, because Creator/Director Anthony DiMieri went to Fordham – Go Rams! – and we take care of our own.)

This little vid is done surprisingly well, with professional production quality and great comedic timing that you might not expect from a bunch of bros. My favorite part, though, has to be post-transformation (bro + hipster = broster) when they all get out at the Bedford stop on the L. LCD Soundsystem’s “North American Scum” plays through a slow-mo of close-ups on jorts, STOP KONY tees, suspenders and Buddy Holly glasses. I died.

Although it’s meant to be a “Girls” parody (the first scene should be familiar to Dunham’s loyal fanbase), it’s also a parody of a few social stereotypes.  The “Bros” demonstrate blatant homophobia and misogyny, while their hip alter-egos talk about basement gin distilleries and warehouse parties. It’s the kind of funny you find when nobody is safe from sarcastic mockery and some playful generalizing. Kudos.

Communication and Media Studies Rules

In Defense of the Comm Major

“Why not spend more time on the computer contemplating the intricacies of my course of study?  It’s not like I have a social life to attend to.”

This is (kind of) what I thought in the process of creating this blog.  I’ve been down this road before, albeit unsuccessfully.  I hope this will be a turn around for me and my exploration into the wonderful world of Communications and Media Studies.

My name is Veronica (cue collective “Hi, Veronica”) and I hope you enjoy yourself on this side of the Internet.

So, Who Do I Think I Am?

  • I am 23 years young, and an undergraduate student of Communications at Fordham University in the Bronx.
  • I live with a graduate student of Communications, and we spend many an evening engaged in debates over McLuhan or Postman or whether Honey Boo Boo is an art work. (If only I were joking.)
  • I’ve done the blogging thing before with a nightlife-oriented start-up in Manhattan.  It was fun, and I like fun, but I’m narrowing my focus these days.
  • I did not become a Communications Major to drink my way through college (that’s what Art History is for).

Before I make this all about me (as I am wont to do) I compel you to ask yourself: What – if any – part of your life is not mediated?

Do you find that your day begins with a get-pumped theme song as your alarm?  Do you watch or listen to or read the news in the early hours of the day?  Do you read textbooks for class or journals and blogs in your professional line of work?  Do you eat lunch in a place with a television?  Do they play music at your local grocer?  Do you send texts and make calls?  Do you e-mail?  Do you use social networking or blog?  Do you read, go to films, or watch favorite television shows in your downtime?

Starting to see my point?

These guys know what I'm talking about.

These guys know what I’m talking about.

I was traveling the interwebs today and found a particularly lovely and simplified version of an argument I have heard over and over again regarding my choice of major:

ObsidianZ writes here:
“I still have no idea what ‘communications’ is supposed to be.
So yeah, I think it’s a bullshit degree.”

I’m not going to make any sweeping generalizations about the other people engaged in my course of study.  I cannot speak for all of us.  Let me, instead, tell you what I have found the butt-of-so-many-jokes “Comm Major” to include and why I do not find it, as *ObZ so eloquently phrases it, “bullshit:”

I became a Communications Major because I wanted a course of study that included provocative, critical analysis about myself and my society, psychology, sociology and philosophy, peppered with popular culture history and contemporary industry training.

I wanted to read and write and learn more about the theories regarding the technology and institutions around which my life revolves.  (What does this mean for me as an American?  A student?  A woman?)

All that, and I wanted a lucrative, fulfilling career at the end of this arduous process commonly known as “higher education.”  Maybe I will go into research and development, academia, television or radio, journalism, music marketing, photography, or public relations.  Maybe I will be able to attain that of which I have dreamed since I learned every letter’s shape – a writing career.

My point is, all this and more is possible with the degree I seek.  If you’re a Communications and Media Studies major like me, welcome friend! If you’re a non-believer, I implore you to read on and keep your mind open. (Or at least, stop calling me a drinking major.)

Until next time,

Twenty-three year old NYC resident, undergraduate at Fordham University, journalist-in-training at WFUV, freelance writer, amateur philosopher, occasional photographer, music-enthusiast, Abe Lincoln fangirl.