It punched to the right. I blocked. Another punch, this time to the left. I blocked again. I quickly jabbed forward, knocking it back slightly. Regaining its balance, it took a step back, swinging its arms. I took the initiative, hitting it with a sharp uppercut to the jaw. Uppercuts seemed to have the best effect, I noticed. It went flying, landing head over heels. Relaxing, I took the HTC Vive virtual reality headset off. I had just demolished a large red battle robot in a one-v-one fight in virtual reality.
That was my experience with Midas Touch Games’ demo, BattleRig, at Augmented Reality World Expo in Silicon Valley. Midas Touch Games is a company that took on one of the largest challenges to VR yet: touchability, or the ability to manipulate and touch objects in virtual reality. It has taken long enough to make a VR headset capable of producing immersive games, videos, and experiences. But the whole idea of VR relies on immersiveness, and if you feel like you are in an unfamiliar physical world, it can make the experience lack credibility. If you touch an object, say, a book on a shelf, and it doesn’t react, that will temporarily take you out of the experience. In an ideal VR experience, everything you touch will react accordingly, therefore limiting the amount of times you actively think “Oh right, this isn’t real. I’m in virtual reality.”
While the ideal virtual reality experience is obviously still years away, Midas Touch Games and many other companies at AWE are helping to make that experience a reality. We have the headsets, we have the 3D models and 360-degree video, now we need to make that content interactive. And Midas Touch Games is doing just that. The one-on-one fight with a robot I was lucky enough to experience was just one example of their unique joint-based physics for VR. Joint-based physics have been in use in VR for inanimate objects for a while, but when it comes to more complex things like, say, robots, it just didn’t work. Midas Touch has created a joint-based system made especially for character-based simulations, what they call the Midas Physics Engine, a 3D engine for virtual reality characters.
While something called “Midas Physics Engine” may sound like PR jargon, it’s quite substantive. I’ve demoed a fair number of VR experiences, and while flying around a prehistoric landscape full of dinosaurs is stunning, the feeling that you are actually there goes away when you fly directly through a stegosaurus. But when you’re petting a dog, and it actually reacts to your touch, and you are even able to pick it up and push it around, as was the demo at the Midas Touch booth, the experience is much more immersive and memorable. And when you’re fighting a robot, when it’s reacting to your punches and vice versa, you really do feel like you’re there, and for me just simply interacting with the characters in your VR experience goes a long way for taking that next step towards the ideal, completely immersive VR experience.
Just two sessions into Augmented Reality World 2016, and the incredible feeling of excitement about the industry had already set in. In a talk entitled The Butterfly Dream: Smart Eyewear in 2031, Dan Eisenhardt discussed the future of AR and the Internet of Things, and it’s no surprise that at a AR/VR/IoT conference the perspective on these technologies is optimistic. Eisenhardt addressed a concern shared by many augmented reality skeptics: will people really wear things on their head all the time?
Basically, this entire conference is a testament to the fact that plenty of large corporations are saying yes. In his talk, Eisenhardt argued that, just as with the book, the wristwatch, and the smartphone, any technology that alters our intake of information, and especially those that alter our appearance even slightly, could suffer backlash. (yes, even books had some backlash, people saying kids were spending too much time inside reading). But just as with the books, the wristwatch, and the smartphone, the utility of these technologies ultimately overcomes these concerns. All the numbers have been pointing in that direction: incredible growth for the last 5 or so years, the colossal projected potential in the industry, and more. But aside from that, to zero in on AR in particular, the potential use-cases are mind-blowing. Walk into a clothing store and a large arrow appears in your vision, identifying clothes in your size? Within a few years, definitely. Texting, calling, searching, helpful information, all showing up in your vision in an unobtrusive and simple way? Almost already here.
While we’re still some distance from having AR glasses look identical to standard prescription glasses, we’re on our way. Even since last year, the glasses being showcased today are smaller, faster, lighter, and have higher-quality displays. The experiences you can already have with headsets like HoloLens and ODG R7 are quite amazing, and it’s only natural that by 2031, fifteen years from now, the best use-cases will have had plenty of time to become mainstream. Eisenhardt went as far as to say the question will not be “should I get AR glasses?” in 2031 but “when will I upgrade my current AR glasses?” In other words, AR will be as ubiquitous as smartphones are today – you can’t leave home without them. I can’t quite decide if that’s scary or exciting.
Sergey Brin, co-founder of Google, wearing the (for the most part) failed Google Glass, the first mainstream AR headset.
I’m a technological optimist; I feel that AR, VR, and other technologies will substantially change the world as we know it. And I expect to jump on the bandwagon as soon as the prices become more competitive. Nevertheless, the notion that our current reality might be replaced with one that is constantly augmented by computer-generated graphics is slightly unsettling. I like reality as it is, and while I trust AR can make our lives easier and more productive, just as smartphones have, I can’t help but wonder whether being completely plugged into an augmented, even completely simulated world with VR would take a certain realness out of life. A realness I enjoy.
Something about AR glasses, wearing computers on our faces, poses a large change to that lifestyle, that reality. Eisenhardt talked about how reality is subjective anyway, how my reality may be different from yours, how easy it is to alter our current perspective of reality, and how holding onto a non-augmented reality is rejecting a technology just because you want to protect something that didn’t really exist in the first place. If you’re fully immersed in a simulation, a VR/AR experience, such that you feel like you’re in another world, move around like you’re in another world, hear, smell, touch, and taste like you’re in another world, who’s to say you’re not in another world?
This is, of course, speculation. By 2031, I would expect AR to be at least something you use on a weekly basis, if not daily. By then, AR would be no different than opening up a laptop, checking your smartphone, or even reading a book. The idea of our current, unaltered, unaugmented reality is slowly being broken down, each new technological advances making our reality seem more and more mutable. AR is very exciting; I’ve had the chance to demo some pretty incredible experiences here at AWE 2016, from beating up a robot and petting a dog with Midas Touch Games, to making music with Subpac, to exploring a prehistoric world of dinosaurs with LifeLiQe, all in AR and VR. And while the novelty of these experiences will soon fade away, they will be replaced by very practical and helpful uses of the technology. While it may be scary, and certainly is exciting, there’s no doubt AR will be a large player in the future of entertainment, enterprise, and more.
Many early adopters of the Internet will remember the chat rooms of the ‘90s, small group messaging “rooms”, almost like group texts on modern smartphones. But these group messaging sites paired you with complete strangers, where people from across the world could discuss whatever they like. Chat rooms on websites like AOL rose to their peak prominence in the 1990s, but then gradually faded in popularity until their eventual demise a decade later.
Some have tried to resurrect the chat room on the modern Internet without much success; the Facebook-bought app Rooms attempted such a resurrection in 2014, but ended up being removed from the App Store a year later in October of 2015 due to a lack of popularity. Chat rooms seem to have fallen out of fashion, other platforms for Internet-wide communication rising to the top such as YouTube, Snapchat and Reddit. And yet, in an interesting moment of irony, Reddit, one of the websites that helped kill off chat rooms, brought them back for a short 8-day social experiment. Called Robin, as in round robin, the experiment was Reddit’s second attempt at leveraging their site’s large user base to learn more about people’s behavior on the Internet, the power of community, and simply how people would react to an interesting chat room mechanism.
For the uninitiated, Reddit can be fairly intimidating, a place where anyone with access to the Internet can interact and discuss any topic they please. While that may not sound all too enticing to you, the way Reddit is organized makes it much more palatable for the average person. Let’s say you’re interested in astronomy and would love to find a community of other astronomy enthusiasts. Just head over to /r/astronomy, the “subreddit” dedicated to everything to do with astronomy and astronomy news. On the astronomy subreddit, or a subreddit on any other topic, Reddit users can post links on that topic and hope they get upvoted to the front page. Every subreddit is a popular meritocracy, comprising primarily of a home page constantly updated with the newest, most popular links and/or discussions posted on that subreddit.
Robin isn’t a subreddit, or a feature on subreddits; actually, it started as just a button (not THE button, last year’s social experiment). When you clicked the button, depicting the outline of a Robin, you are brought to a page with one more button, this time saying “participate”. Upon clicking that button, you were brought into a chat room with you and one other random Reddit user. You and that Reddit user could chat for 1 minute about whatever you would like, and then you would have to vote on one of three options in the top right-hand corner of the page: stay, grow, or abandon.
Think of Robin as creating a village. If the majority of you and your randomly-assigned village-mates vote grow, then your “village” is merged with one of the same size. If the majority chooses stay, you build a wall around your village and plan to stay there for “life,” meaning you get a private subreddit made for that group that only the group members can join. And finally, if the majority votes abandon, the village is immediately closed and you can start again from the beginning, and if a couple of individuals vote to abandon then they alone are kicked out of the village. (in case you didn’t catch on, village means chat room in this analogy)
This would seem like a pointless voting mechanism if not for the fact that Robin in based on a chat room, a place where anyone can talk to anyone else in the world. Every time your group grows, the time you get to talk to your fellow “village-mates”, from 1 minute with 2 people, 3 with 4 people, 7 with 8 people, 15 with 16 people, and 30 with any higher number of participants. This means that, despite what you may think of people on the Internet, you can actually have interesting conversations. In my experience, as I spent a fair amount of time playing Robin when it was still running from April 1st to April 8th, the conversations you have change topic very fast and yet are still very entertaining and fun.
On top of that, the very mechanism of growing, staying, or abandoning sparks conversation, users campaigning for different votes. In one of my chats, I was part of an “ABANDON2016” campaign, as an overwhelming number of memes had flooded our chat. We were successful, to the dismay of the meme makers. In fact, for the people who vote stay, it’s not actually because they want a private subreddit with a bunch of random people who share most likely no interests; it gets boring after a day.
The conversations you have getting people to stay is the fun part, the part that makes it all worth it. The reason Robin was more than just a chat room was because it gave every single member of the chat a sense of attachment, to the people in the chat, and to the chat itself. Because it takes 30 minutes to get to the voting point of a 16 person chat, you have spent quite a while on any certain chat. Also, because of the way the chats merge, everyone feels like they had a part in starting what would later become a forum for global communication. In only a couple chats, I talked to people, usually men in their 20s, from places such as Turkey, the Netherlands, England, Israel, and more. This is the reception I got when I announced I had to leave a chat room, after participating in the election for 15-20 minutes prior:
(my username is aaron21morgan, and the bottom half of the running chat is mostly my fellow chat-mates’ response to my parting announcement)
In the end, after about a week up and running, Robin shut down just as The Button did a year ago. The largest chat ended up having 5,000+ people, and was apparently so big that the Reddit servers were having trouble just keeping it afloat. Now that the fun is over, we can look back on Robin as an interesting experiment and an entertaining game. We don’t often think about the fact that the Internet really does connect everyone in the world, and just the act of talking to a handful of them in a chat room can make that connection feel all the more real. Robin may have been a stupid Internet chat room to outsiders, but to those participated it helped humanize the average Internet commenter we encounter on an everyday basis. It may just be me, but having a simple conversation with someone thousands of miles away, even just for a couple of minutes, brings back the sense of wonder about the capabilities of the Information Age from back when the Internet was first being brought into mainstream culture.
This article is the second in a two-part series on the advertisement conundrum of online video. If you haven’t already, check out the first article in the series HERE.
In the previous article in this series, I explained the current monetization situation in online video, and how the $2 CPM ($2 per thousand views) is, as of right now, the industry standard ad rate. As I have begun to learn, along with many others in the field of online video, the $2 CPM is neither a sustainable nor beneficial model to anyone but the companies paying for the ads. There are a few reasons why the current $2 CPM standard simply doesn’t work. First of all, it’s just too low; with only $2 per thousand views, YouTubers have to grow a very large audience, which is incredibly hard in the current YouTube landscape. There has been an explosion of creators and content using YouTube in recent years, and as of now, it’s very unlikely that any one small YouTuber will hit it big, simply because of the enormity of the amount of channels out there in the first place. Secondly, even once you factor in the $2 ad rate, YouTube takes away 45% of that revenue for their own profit. While it isn’t surprising that YouTube has a “platform tax”, after all, Google is a publicly-traded company whose investors care about the bottom line. Yet 45% can be a big hit to smaller and larger channels, and compares unfavorably to Apple’s already high 30% revenue share for App Store products. But my suspicion is that this won’t be the part of the equation that will be changing, as YouTube has in the past addressed this issue and said nothing about reducing their share of the monetization.
Finally, the $2 CPM restricts the production quality of growing YouTubers who are forced to produce cheaper content more often to make ends meet. YouTube has addressed this as well, and their counterargument is that YouTube isn’t and wasn’t built to be the main source of revenue for a channel, and only really works in that capacity for smaller channels. And it’s true, YouTube is good at growing brands and personalities, although that attribute is shrinking because of the low odds of being successful.
Basically, they’re suggesting that you as a YouTuber derive your main or secondary source of income elsewhere, perhaps through sponsorships and other marketing deals. And while many channels do that type of marketing, at some point creators just don’t want to surrender their editorial control in exchange for money, most likely part of the reason they came to the platform in the first place. Hank Green, in his Medium article on the subject of the $2 CPM, posed the “just ask” strategy, i.e., just ask your audience for money. You might think that strategy would never work, but the website Patreon is built directly off of that strategy, and it has grown substantially and has become a great source of money for many creators.
Creators tell their audience that they would like, in exchange for some small perks, those who can afford and would like to, “pledge” to give them a certain amount of money regularly, commonly every video or every month. This allows creators to realize another source of income, without such a prominent middle man, and simply because the viewers like their content and want to creator to make more and better videos.
You may be skeptical that this strategy would work, but take Crash Course, which earns $29,000/month directly through Patreon, and they aren’t the only ones. As Hank Green wrote in his article, giving to creators because you like their work and want them to make better content doesn’t encourage more lower quality videos but “it encourages a different kind of content. Instead of challenging creators to figure out how to get the highest view counts, creators have to puzzle out how to make the most valuable content.”
Even still, we can’t expect Patreon to completely cover a creator’s budget, as even for large YouTubers only a small percentage of your audience is willing and able to pay on a regular basis. Ads certainly won’t cover the costs, and while sponsorships hopefully would fill the rest of the budget, for many creators you still end up on the short side of where you want to be. YouTube has only been around for a little more than 10 years, but it’s still on a path to realizing its full potential in terms of influence and popularity. Hank Green feels that the rough transition from TV to online video may actually be a good thing for creators:
“Part of me is actually glad that advertisers have been so slow to adapt. It’s made it clear to a lot of people that advertisements aren’t the only (or even the best) way to monetize content… What that leads to isn’t just new business models — we’re seeing new creation models, new audience relationships, and new kinds of content. With a couple of simple new tools, the economic arrow is suddenly pointing in new directions, and I’m very excited to see where it leads us.”
We’ll see where the ad conundrum of YouTube takes us, and if it can be solved in the near future. Right now, YouTube ad rates are valued at 32x less than TV and 250x less than film, figures that I doubt anyone who is familiar with the platform would agree with. Money guides everything, even an open platform with a negligibly small barrier-to-entry. In my opinion, despite all the constraints that are financially-cornering creators, the community as a whole will get past this issue and grow a thriving, profitable, and revolutionary (well, more revolutionary than it already has been) platform for creators and audiences of all kinds.
With the possible exception of the Super Bowl, I’d bet it’s safe to say that nobody likes ads. Whether before a movie or video, in commercial breaks during television programs, or in the middle of your favorite podcast, nobody really enjoys being told to buy this product or use this service (often in a cringe-worth way) while they are enjoying their entertainment. Yet advertisements aren’t going away anytime soon; with the larger and larger audiences their ads are reaching, companies remain willing to allocate precious dollars to get their name out in every way they can. In the world of Internet publishing, ads have persisted as the staple of a creator’s income, despite significant shifts in the media landscape. But for online video, currently dominated by YouTube. advertisements have been a challenged revenue channel for creators hoping to earn a living.
I love YouTube and have a massive respect for the creators who have made it their full-time occupation to publish videos on the platform. These individuals spent an incredible amount of time and effort to become popular enough just to quit their day jobs and spend their time earning a living via YouTube. The sad part is that making a living on YouTube is harder than one might think. With popular YouTubers like PewDiePie making up to $7 million per year, it might be easy to regard YouTube as an easy path to fame and riches. But really, every YouTuber with even just 5,000 subscribers have put their heart and soul into their videos. As it is, money coming from ads just isn’t enough to allow YouTubers to start making videos full-time until they become very popular, a level which many never reach.
Let’s do the math. The average personal income in the United States is roughly $30,000. The current YouTube ad rate is a $2 CPM ($2 for every thousand views). To earn even the average U.S. income, a YouTuber creating weekly videos (a common schedule) would have to average nearly 300,000 viewers per video, an average usually only met by a YouTuber with around 2 million subscribers. (this varies from channel to channel) Of course, the rate at which you create videos is key in this calculation; if you make a video every day, the average view count drops to a more plausible 40,000. Compare that to the average CPM rate for TV, which is $19 (for an average 22-minute show). With that rate, you would only have to get 30,000 views per weekly video to reach the national average – much more sustainable.
This isn’t just about YouTubers making more money because their media peers in television and film earn more. I’m not writing this out of pity for the struggling YouTubers who can’t earn a living wage yet are spending all their time trying to grow their audience. The reason the $2 CPM needs to be increased is because it simply isn’t enough to allow YouTubers to grow and make the great content we all want to watch. Take Olga Kay, a YouTuber with around 1 million subscribers across her five channels. In an article written in the New York Times, Olga talked about her hectic work schedule and about how “If we [her friends] were coming to YouTube today, it would be too hard. We couldn’t do it.”
Olga said in the article that she has made $100,000 to $130,000 every year for the last three years, which is a good income; yet she is still constantly stressed about finances, as much of that $100,000 goes straight back into her channels to pay for editors, equipment, etc. Let’s be honest: no one making twenty videos a week, almost three per day, especially with 1 million subscribers, should be that worried about finances.
This is the first part in Fast Forward’s two-part series on the YouTube’s advertisement and monetization conundrum. Stay tuned for the second article in the series over the next few days!
Today is the first day of 2016, but before embracing the New Year here at Fast Forward, it’s time to look back upon the past year in smartphone innovation with our third annual App of the Year award. Without further adieu, I present to you the top 5 apps of 2015:
Without a doubt, Drafts is one of the most useful, if not the most useful, apps on my phone right now. I’ve used Drafts for a couple of years now and it’s still one of my most used productivity tools on the iPhone. Like a notes app on steroids, Drafts allows you to easily and quickly write down notes and share the with a click of a button. With customizable options, you can share your note to any one of Drafts’ long list of options, from Twitter to Evernote to Dropbox to Email to Reminders et cetera, et cetera. You get my point. Drafts has become my personal hub for ideas, notes, and frankly writing anything that I happen to want to write down: phone numbers, article ideas, shopping lists, and more. Plus, with a new design and an update that includes some very helpful shortcuts using Apple’s new force touch, Drafts is becoming faster and better with developer updates that leverage Apple’s new features. At least for me, the usefulness of the app is incredible; while I often simply send myself emails with reminders and notes, in Drafts this task is simple while also offering the option to share or save my notes however I like.
Although podcasting has been around since the 1980s, then known as “audioblogging”, podcasts only really started to pick up steam in the last 5 years. With the first season of the hit podcast Serial in 2014, the medium finally became a large part of modern culture, as Serial’s influence was comparable to a popular weekly TV series. In fact, in it’s first season, Serial was averaging an incredible 1.5 million downloads per episode, a number certainly qualifying the show as significant. Podcasting in general has grown as well, studies showing that approximately 39 million Americans have listened to a podcast over the last month. So if you aren’t already listening to a bunch of podcasts, as I am, it might be time to start. There are podcasts for everyone, in topics ranging from tech to comedy to sports, and Casts is a great app to help you find new podcasts and listen to your favorites.
While other podcast apps dominate the market, including Apple’s own version, I personally use Casts for its simplicity and aesthetically pleasing design. Once you’ve found your favorite podcasts, which you “subscribe” to, the icons of all your podcasts will show up in a nice, tiled design on your homepage, making it easy for you to navigate and download the ones you want to listen to. With helpful navigation bars and intuitive design elements, Casts is, in my opinion, the best podcasting app out there.
2015 has been another great year for the social media giant Instagram. With over 400 million active users, 20% of the users of the Internet in general, Instagram has grown from a small trendy platform to the ubiquitous giant social media app it is today. In the past year, besides continuing to be the favorite social media of many youth around the world, Instagram has made many interesting and, in my opinion, good changes. First of all, they dropped the “square only” rule, adding functionality to post photos of all sizes. This allowed people much more freedom in what they post, as often regularly proportioned images wouldn’t fit in Instagram’s odd square constraint. Along with that, Instagram, like Drafts, is easily one of the apps that has best used the iPhone 6S’ new feature, “Force Touch.” Just by force touching on usernames, photos, and other parts of the interface, you can easily get a preview of whatever you touch on, which can actually save a lot of time when browsing the new explore section or sliding through your feed. In sum, while Instagram has stayed primarily the same, a couple of changes made during 2015 made the app that much better than it already was, which was pretty dang good.
Periscope was, in my view, the breakout app of the year. First being bought by Twitter, then winning Apple’s App of the Year award, Periscope emerged as the clear winner in the tough rivalry that emerged with competitor Meerkat. Both live video streaming apps, Periscope triumphed as the leading app in that space, with more users (10 million to Meerkat’s 2 million), a better design, at least in my opinion, and the backing and integration with Twitter. In the app, you can watch live streams from all over the world, and even only about 10 months since it’s launch, people have already started using the platform in a wide variety of interesting and fun-to-watch ways. (Here’s one of my favorite examples) I’m sure that over the next couple years streaming apps will grow into a whole new medium, and Periscope will be leading that charge.
Fitness has always been a large part of the smartphone/wearables market. Exercise is just something that everyone knows they should do, and Runkeeper helps us stick to our goals by track our runs, bike rides, and other activities with the iPhone’s GPS. Exercise apps should compile a host of stats, have an intuitive design, be easily navigable (as you don’t want to be fumbling around with your phone while you are running or biking), and just be fun and enjoyable to use. Runkeeper is all of that and more. The main screen is very easy to use, which helps when exercising, and once you’re done the app will give you your time, calories burned, distance traveled, average speed, and more. You can set goals, connect with friends, play music, and pretty much everything your would want out of a running app. With almost 5 stars on the app store, Runkeeper has risen to prominence in the fitness section over the last year.
It’s 1916. A relatively unknown (see what I did there) physicist named Albert Einstein has just released a couple of papers about his very weird but interesting Theory of Relativity, and you can’t decide what you want to think about it. The papers have gained Einstein some popularity in the physics community, many agreeing with what he had to say, some disagreeing. Many years later, once the theory was proven through observations of Mercury’s orbit and other various tests throughout the years, Einstein became an internationally renowned, scientific rock star. Now it’s 2015, 100 years after Einstein introduced his ingenious theory. Einstein is still known as the epitome of human intelligence, and his fame and prestige is as widespread as ever, while the repercussions of his work still permeate the modern scientific landscape – so much so that it allows us to predict the future.
Before we get into the whole time travel stuff, it might help to understand more about what exactly General Relativity is. Despite the fact that Einstein is known as a genius worldwide, the specifics of what his theory actually means are lost on most people who haven’t extensively studied it. Whereas we see time and space as distinct, Einstein thought of them as one thing: “spacetime”. (Try not to think too hard about that merging space and time part.) Where Newton thought that there was this mysterious force called gravity that radiated from anything with mass, Einstein had his own theory, which assumed that, rather than using an enigmatic force to pull nearby things to it, objects with mass (especially large ones) curve spacetime. So, when following the easiest path through spacetime, which all objects naturally do, objects near another large object, say the Earth, naturally get pulled toward it. Now there’s a proposition for you: just by existing, you disrupt the fabric of the universe by curving it.
But it’s true, and as I wrote earlier, general relativity has been confirmed through many different experimental observations over the years. One such confirmation came through the discovery of a phenomenon that Einstein predicted in his papers, but never expected to be observed because of its rarity: gravitational lensing. Gravitational lensing occurs around large objects, though not just large as in the Earth or even the Sun, but large as in supermassive galaxy clusters. As it turns out, Einstein’s idea that things with mass curve spacetime not only explains gravity but also implies gravitational lensing. Because large objects curve spacetime, really large objects like galaxy clusters curve spacetime much more, enough so that light coming from behind the cluster becomes visibly curved.
The light coming from an object directly behind the galaxy cluster, a smaller galaxy cluster for instance, will get curved around the cluster in the foreground and scattered away from its previous course, potentially toward the awaiting telescopes here on Earth. Einstein thought it would be unlikely that such a situation that we could observe on Earth would happen anytime soon; that a gravitationally lensed light source would happen in our direction, that the light would be lensed toward us, and that we’d be ready to receive it. He happened to be wrong on this one, as we have observed many instances of gravitational lensing in action over the years, the image below being one such example.
And one more thing: gravitational lensing essentially peers into the past. Because light takes time to travel to Earth, the farther away you look in a telescope, the further back in time you see. This doesn’t affect us in day to day life, but with large space telescopes like Hubble we can see billions of year into the past. Since gravitational lensing acts very similar to a telescope, taking light from far away and magnifying it, the foreground galaxy cluster acting as the lens, we can legitimately say that gravitational lensing peers into the past.
So, aside from being a fascinating phenomenon and an example of General Relativity in action, gravitational lensing has been used over the years be a tool to peer into the universe’s history, and recently, into the future as well. Researchers at the University of California, Berkley, published a paper in Science that detailed their observations and subsequent research on a supernova they named Refsdal after Norwegian astronomer Sjur Refsdal. Patrick Kelly, an astronomer at the university, noticed something odd while looking through photos taken by the Hubble Space Telescope. What Kelly noticed was what’s called an Einsteinian Cross, such as the one below:
An Einsteinian Cross occurs when gravitational lensing bends the light around a galaxy cluster and toward Earth in a way that projects 4 images of the same star around the forefront galaxy in the shape of a cross. In this case, the star was actually a supernova, when a star explodes in a fantastic burst of energy and light. Supernovae are one of the biggest and brightest events in the universe, releasing up to around 1044 joules, or the total energy released by an average star in it’s 10 billion year long lifetime. But supernovae are fairly rare, so you can imagine how lucky it was to observe a gravitationally lensed one.
As it turns out, Refsdal is actually 9 billion light years away, and since the supernovae is a one-time event, we can conclude that gravitational lensing has allowed us to photograph the supernovae 9 billion years in the past, almost three-quarters of the way to the beginning of the universe. Because of an odd phenomenon in spacetime, and lucky happenstance, we get to observe the death of a star billions of years ago. And not only that, but we get to see it multiple times.
While in most drawings and explanations of gravitational lensing, we see the light being bent around a perfectly spherical bubble of gravitational attraction, whereas in reality that gravitational bubble can be warped. Essentially, the path around the cluster can differentiate the light paths enough so that some take longer than others. The direct effect of this differentiation is that we get to see the supernova over and over again, as the light paths keep hitting us at different times. We only observed one so far, in November of this year, but we can assume that we missed the explosion being replayed twice in the 20th century, and we can even predict that another will appear in the next 1-4 years.
And there we have it! We have successfully predicted the future, although admittedly in a fairly specific way. But the usefulness of this situation spans more than just the coolness of predicting the future and seeing into the past. “It’s a wonderful discovery,” said Alex Filippenko, a professor of astronomy at Berkley and member of Kelly’s team. “We’ve been searching for a strongly lensed supernova for 50 years, and now we’ve found one. Besides being really cool, it should provide a lot of astrophysically important information.” For instance, with more research the scientists may be able to gain insight into how supernovae work, how gravitational lensing work, in turn the nature of spacetime, and even how dark matter works, which makes up a large portion of the gravitational pull affecting the light from the supernova’s original explosion. Discoveries like these are not only exciting and interesting, but they allow us to continually understand more and more about the world, and universe, around us.
Just this week, China issued a “red alert” for the country’s capital, Beijing. The alert was used as a warning to their population that the smog levels in Beijing and the surrounding area has risen to a dangerous level, that is, even more dangerous than before. In response, schools are advised to be shut down, the number of vehicles on the roads will be restricted, and all factories and construction sites will be temporarily closed. And it’s not as if no one could see this coming, as China’s pollution problem has only worsened in recent years. Wearing masks over your mouth, like the type you are given at the doctor’s office when you are sick, has become commonplace all over the city, with some people even donning large filtering masks instead. In fact, companies are now taking advantage of this demand, releasing a range of masks in different colors and styles. Beijing has reached a point where the pollution is inhibiting daily life in a major way: outside of buildings, the ever-present smog enveloping the city means that simply walking outside has come dangerous.
Although it’s clear from a glance at the smog that Beijing’s pollution situation is more than just unhealthy, it’s always good to check that with science. An AQI, or Air Quality Index, is the way that most countries around the world measure the level of pollution in a simple way that the public can understand and act upon. In China, the AQI is measured very frequently throughout major cities and often influences behavior, e.g., whether they go outside at a given time. The recommended, healthy AQI level globally is from 0-50. Here in the Bay Area, where I live, the AQI is around 15. In the city of San Francisco it rises to around 40, which is still good. In London the AQI is around 75, which is worse but still considered acceptable. The exact hierarchy of the severity of the pollution varies between countries, but both China and the United States use a system that is very similar to the table below:
At the time of writing, the AQI level in Beijing was 323. Less polluted areas of the city have an AQI of 220, worse areas an AQI of a whopping 445, just 55 points shy of literally being off the scale (typically AQI is 0-500). Clearly, China have dug themselves into a whole that seems very difficult to escape. “I’m used to the smog,” Wolf Hu, a resident of Beijing, told CNN. “I’d find a day when the sky is blue unusual.”
We can all agree, even the Chinese government, that the air in China’s major cities is extremely hazardous and puts their population in unnecessary danger and discomfort. But recent studies have shown that the incredibly high pollution levels in China do much more than put their population in discomfort. The most dangerous pollutant, PM2.5, where PM stands for particulate matter, targets particulates that are 2.5 microns in diameter or less. For comparison, a human hair is around 40-50 microns across. Particular matter consists of dust, soot, smoke, liquid droplets, or anything that happens to be floating around in the air at the time. While particulate matter is incredibly small, if any happens to float into your lungs when you breathe, the particles can cause a lot of damage, including increasing risk of a variety of heart and lung related diseases.
It should be no surprise then, as the PM2.5 level all over the country is very high, that the population have started to feel the consequences. Using new data released by the Chinese government, a study done by researchers at the University of California, Berkeley found that 1.6 million people die from heart, lung, and stroke problems stemming from breathing pollution in China every year. That’s 4,000 people dying every day, 17% of all of the annual deaths in the country. The very same study also found another eye-opening statistic: breathing in the pollution in Beijing and other cities in China does equivalent damage to your lungs as smoking 40 cigarettes per day does.
The state of China’s air is, to say the least, abysmal. The term “airpocalypse” has already been coined to refer to the problem. To be honest, it like the setting of a young-adult dystopian novel than the state of one of the global superpowers of the 21st century. To their credit, the Chinese government has (belatedly) recognized the problem. Li Keqiang, the Chinese Premier, has stated that China must “fight with all their might” against pollution, and has previously “declared war” on pollution, although no real plan or action has been taken to reduce pollution or cut emissions. China and the United States created a joint agreement for China to slow, peak, then cut carbon emissions by 2030, and both nations attended the landmark Paris climate summit in which leaders from around the world are collaborating on a solution to limit climate change. We may finally be going in the right direction, but it will certainly take an effort from the entire world to try and keep the world a clean, vibrant, and vivacious place, free of carbon. If there’s one thing we can all agree on, it’s that a world where a blue sky is rare and the outside is feared is no place that anyone wants to live in.
Sources:http://www.cnbc.com/2015/08/18/china-air-pollution-far-worse-than-thought-study.html http://www.theguardian.com/environment/2015/dec/07/beijing-pollution-red-alert-smog-engulfs-capital http://www.theguardian.com/cities/2014/dec/16/beijing-airpocalypse-city-almost-uninhabitable-pollution-china http://aqicn.org/city/beijing/ http://www3.epa.gov/pm/ http://www.cnn.com/2015/12/07/asia/china-beijing-pollution-red-alert/ http://www.wikiwand.com/en/Air_quality_index#/United_States http://www.theguardian.com/environment/2015/mar/05/china-vows-to-fight-pollution-with-all-our-might
You’re driving down the road in your Toyota Camry one morning on your way to work. You’ve been driving for 15 years now and pride yourself on the fact that you’ve never had a single accident. And you have to drive a lot, too; every morning you commute an hour up to San Francisco to your office. You pull into a two-lane street lined on both sides with suburban housing, and suddenly realize you took a wrong turn. You quickly look down at your smartphone, which is running Google Maps, to find a new route to the highway. When you look back up, you’re surprised to see a group of 5 people, 3 adults and 2 kids, have unknowingly walked into your path. By the time you or the group notice each other it’s too late to hit the break or for the pedestrians to run out of the way. Your only option to save the 5 people from being injured, or even killed, by your car is to swerve out of the way… right into the path of a woman walking her child in a stroller. You notice all of this in the half a second it takes you to close the distance between you and the group to only 3-4 yards.
You now have but milliseconds to decide what path to take. What do you do? But more to the point of this article, what would an autonomous car do?
That narrative is a variant of the classic situation known as the Trolley Problem. The Trolley Problem has many variations, some more famous than others, but all of them follow the same general storyline: you must choose between accidentally killing 5 people (e.g., hitting them with your car) or purposefully making an action (e.g., swerving out of the way) that kills one person. This type of situation is obviously one that no one wants to find themselves in, and is so unlikely that most people avoid it their entire life. But in the slim cases where this situation occurs, the split-second decision a human makes will vary from person to person and from situation to situation.
But no matter the outcome of the tragic event, if it does end up happening, the end result will be generally be the fault of a distracted driver. What will happen, though, when this decision is completely in the hands of an algorithm, as it will be when autonomous cars ubiquitously roam the streets years from now. Every new day autonomous cars become more and more something of the present rather than the future, and that leaves many worried. Driving has been ingrained in us for century, and for many, giving that control up to a computer will be frightening. This is despite the fact that in the years that autonomous cars have been on the roads, their safety record has been excellent, with only 14 accidents and no serious injuries. While 14 may seem like a lot, keep in mind that each and every incident was actually the result of human error by another car, many of which were the result of distracted driving.
I’d say that people are more worried about situations like the Trolley Problem, rather than the safety of the car itself, when driving in an autonomous car. Autonomous cars are just motorized vehicles driven by algorithms, or intricate math equations that can be written to make decisions. When an algorithm written to make a car change lanes and parallel park has to make almost ethically impossible decisions, choosing between just letting 5 people die or purposely killing 1 person, we can’t really predict what it would do. That’s why autonomous car makers can’t just let this problem go, and have to delve into the realm of philosophy and make an ethics setting in their algorithms.
This won’t be an easy task, and will require everyone, from the car makers to the customers, thinking about what split-second decision they would make, so they can then program the cars to do the same. This ethics setting would have to work in all situations; for instance, what would it do if instead of 5 people versus one person, it was a small child versus hitting an oncoming car? One suggested solution would be to have adjustable ethics setting, where the customer gets to choose whether they would put their own life over a child’s, or to kill one person over letting 5 people die, etc. This would redirect the blame back to the consumer, giving him or her control over such ethical choices. Still, that kind of a decision, which very well could determine fate of you and some random strangers, is one that nobody wants to make. I certainly couldn’t get out of bed and drive to work knowing that a decision I made could kill someone, and I’d bet I’m not alone on that one. In fact, people may even avoid purchasing an autonomous car with an adjustable ethics setting just because they don’t want to make that decision or live with the consequences.
So what do we do? Nobody seems to want to make kind of decisions, even though it is absolutely necessary. Jean-Francois Bonnefon, at the Toulouse School of Economics in France, and his colleagues conducted a study that may help us all with coming up with an acceptable ethics setting. Bonnefon’s logic was that people will be most happy with driving a car that has an ethics setting close to what they believed is a good setting, so he tried to gauge public opinion. By asking several hundred workers at Amazon’s Mechanical Turks artificial intelligence lab a series of questions regarding the Trolley Problem and autonomous cars, he came up with a general public opinion of the dilemma: minimize losses. In all circumstances, choose the option in which the least amount of people are injured or killed; a sort of utilitarian autonomous car, as Bunnefon describes it. But, with continued questioning, Bunnefon came to this conclusion:
“[Participants] were not as confident that autonomous vehicles would be programmed that way in reality—and for a good reason: they actually wished others to cruise in utilitarian autonomous vehicles, more than they wanted to buy utilitarian autonomous vehicles themselves.”
Essentially, people would like other people to drive these utilitarian cars, but less enthusiastic about driving one themselves. Logically, this is a sensible conclusion. We all know that we should make the right decision and sacrifice your life over that of someone younger, like a child, or a group of 3 or 4 people, but when it comes down to it only the bravest among us are willing to do so. While these scenarios are far and few between, the decisions made by the algorithm in that sliver or a second could be the difference between the death of an unlucky passenger or an even more unlucky passerby. This “ethics setting” dilemma is a problem that can’t just be delegated to the engineers at Tesla or Google or BMW; it has to be one that we all think about, and make a collective decision for that will hopefully make the future of transportation a little more morally bearable.
Earlier this month at Apple’s annual product event, a new device was released that caught some by surprise. Historically, Apple isn’t a company that’s known for experimenting with different product lines, sizes, colors, or software designs. In recent years, Apple has started to branch out from their traditional iPhone and iPad lines, and brought the iPhone 5C, the line of iPhone Pluses, the iPad Minis, Apple TVs, and more. Clearly, they are trying to produce more options for their customers to choose from when buying one of their phones or tablets, which if anything benefits the customer more than Apple itself. But at the September 9th conference, Apple announced a product that baffled techies and average consumers alike: the iPad Pro.
Like the positioning of the Macbook Pro and the Mac Pro, the iPad Pro is essentially just a higher performance iPad. The specs for the device are promising; most importantly, the resolution is better than the high-end Macbooks, at 264 pixels per inch, even beating out it’s newfound competitor, the Surface Pro. It has a 10-hour battery life, which is fairly good for the device’s size, and again beats out the Surface at 9 hours. But of course, the one spec that surprised everyone was the size: the iPad Pro has an insane 12.9-inch screen.
That’s 3.2 inches bigger than the recently released iPad Air 2, the latest installment in the iPad line. Now, while the product had been rumored for months, far fewer could have expected a super-sized iPad over a year ago, in part because, unlike the upsized iPhones few really saw the need for a giant iPad. The iPad Air 2 is already a pretty good size at 9.7 inches, and adding three inches on to that doesn’t really rectify the $300 price increase. Sure, as Apple has been partly marketing the device, the iPad Pro would be a great device for watching and consuming media, like watching movies, readings articles, and perhaps even playing games. The iPad Pro could easily replace your laptop as your main entertainment consumption device, although personally, I would just spend the extra $200 to get the Macbook Air of the same size, as certain functionalities of the Macs over iPads are important to me and my work. And to be honest, the iPad line is starting to feel a little like this:
That’s not to say that the iPad Pro is a bad addition to the iPad line. If anything, the addition to the product line helps make the iPad line a better fit for consumers or professionals with specific use cases. For instance, one example that everyone came up with simultaneously after the iPad Pro’s release was for artists. The iPad Pro is a great size for a digital art pad, and the excellent display only makes adds to the use case. This hypothesis, that Apple was targeting artists, was only reinforced by their release of new product, an almost parody-esque product, the Apple Pencil.
Apple Pencil is, as you probably guessed, a stylus. Designed to work with the iPad Pro, Apple Pencil is Apple’s attempt at getting into the stylus market, although the Pencil may only work with the iPad Pro. The stylus is actually a very good stylus; it has a very good response time when in use, is pressure sensitive, and overall has a very fluid and smooth feel to it. That’s all and well, and will definitely help out artists when using the iPad Pro, but the reason that this device is so surprising is because of Steve Jobs’ views on the product category. Although Jobs isn’t around to keep Apple going anymore, he did have some opinions that surely shaped the way Apple progressed each year, and this is the first time we’ve seen evidence of Apple disregarding what he thought. Jobs has a strong opinion against styluses, expressing how he thought they were cumbersome and hard to keep track of. “Who want’s a stylus?” he said in 2010 keynote speech. “If you see a stylus, they blew it.”
The Apple Pencil aside, the iPad Pro is an interesting product. No doubt it’s a high-quality device; it has great specs and the big screen just makes it a great content viewing platform. But for 800$? I certainly wouldn’t spend the money, but for people who can (or just want to) it’s a great purchase, as long as they know why they want it. If you’re just looking for an iPad, the iPad Air 2 is a great choice. But the iPad Pro is the kind of niche device that’s great for the people who have a reason to use it, such as artists, but maybe not as profitable for the company, and certainly not the type of device you expect Apple to release. Still, that may show they are trying to branch out into more product types and categories, which very well may lead to some great products in the future.