{"id":212260,"date":"2017-08-18T05:01:07","date_gmt":"2017-08-18T09:01:07","guid":{"rendered":"http:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/instagram-ceo-kevin-systrom-on-free-speech-artificial-intelligence-and-internet-addiction-wired\/"},"modified":"2017-08-18T05:01:07","modified_gmt":"2017-08-18T09:01:07","slug":"instagram-ceo-kevin-systrom-on-free-speech-artificial-intelligence-and-internet-addiction-wired","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/free-speech\/instagram-ceo-kevin-systrom-on-free-speech-artificial-intelligence-and-internet-addiction-wired\/","title":{"rendered":"Instagram CEO Kevin Systrom on Free Speech, Artificial Intelligence, and Internet Addiction. &#8211; WIRED"},"content":{"rendered":"<p><p>Skip Article Header. Skip to: Start of  Article.  <\/p>\n<p>    I sat down with Kevin Systrom, the CEO of Instagram, in June to    interview him for my feature story, Instagrams CEO Wants to Clean Up the    Internet, and for Is Instagram Going Too Far to Protect Our    Feelings, a special that ran on CBS this week.  <\/p>\n<p>    It was a long conversation, but here is a 20-minute overview in    which Systrom talks about the artificial intelligence Instagram    has been developing to filter out toxic comments before you    even see them. He also discusses free speech, the possibility    of Instagram becoming too bland, and whether the platform can    be considered addictive. Our conversation    occurred shortly before Instagram introduced the AI to the public.  <\/p>\n<p>    A transcript of the conversation follows.  <\/p>\n<p>    Nicholas Thompson, Editor-in-Chief: Morning,    Kevin  <\/p>\n<p>    Kevin Systrom, CEO of Instagram: Morning! How    are you?  <\/p>\n<p>    NT: Doing great. So what I want to do in this    story is I want to get into the specifics of the new product    launch and the new things youre doing and the stuff thats    coming out right now and the machine learning. But I also want    to tie it to a broader story about Instagram, and how you    decided to prioritize niceness and how it became such a big    thing for you and how you reoriented the whole company. So Im    gonna ask you some questions about the specific products and    then some bigger questions  <\/p>\n<p>    KS: Im down.  <\/p>\n<p>    NT: All right so lets start at the beginning.    I know that from the very beginning you cared a lot about    comments. You cared a lot about niceness and, in fact, you and    your co-founder Mike Krieger would go in early on and delete    comments yourself. Tell me about that.  <\/p>\n<p>    KS: Yeah. Not only would we delete comments    but we did the unthinkable: We actually removed accounts that    were being not so nice to people.  <\/p>\n<p>    NT: So for example, whom?  <\/p>\n<p>    KS: Yeah well I dont remember exactly whom,    but the back story is my wife is one of the nicest people    youll ever meet. And that bleeds over to me and I try to model    it. So, when we were starting the app, we watched this video,    basically how to start a company. And it was by this guy who    started the LOLCats meme and he basically said, To form a    community you need to do something, and he called it Prune    the trolls. And Nicole would always joke with me, shes like,    Hey listen, when your community is getting rough, you gotta    prune the trolls. And thats something she still says to me    today to remind me of the importance of community, but also how    important it is to be nice. So back in the day we would go in    and if people were mistreating people, wed just remove their    accounts. I think that set an early tone for the community to    be nice and be welcoming.  <\/p>\n<p>    NT: But whats interesting is that this is    2010, and 2010 is a moment where a lot of people are talking    about free speech and the internet, and Twitters role in the    Iranian revolution. So it was a moment where free speech was    actually valued on the internet, probably more than it is now.    How did you end up being more in the prune the trolls camp?  <\/p>\n<p>    KS: Well theres an age-old debate between    free speechwhat is the limit of free speech, and is it free    speech to just be mean to someone? And I think if you look at    the history of the law around free speech, youll find that    generally theres a line where you dont want to cross because    youre starting to be aggressive or be mean or racist. And you    get to a point where you wanna make sure that in a closed    community thats trying to grow and thrive, you make sure that    you actually optimize for overall free speech. So if I dont    feel like I can be myself, if I dont feel like I can express    myself because if I do that, I will get attacked, thats not a    community we want to create. So we just decided to be on the    side of making sure that we optimized for speech that was    expressive and felt like you had the freedom to be yourself.  <\/p>\n<p>    NT: So, one of the foundational decisions at    Instagram that helped make it nicer than some of your peers,    was the decision to not allow re-sharing, and to not allow    something that I put out there to be kind of appropriated by    someone else and sent out into the world by someone else. How    was that decision made and were there other foundational design    and product decisions that were made because of niceness?  <\/p>\n<p>    KS: We debate the re-share thing a lot.    Because obviously people love the idea of re-sharing content    that they find. Instagram is full of awesome stuff. In fact,    one of the main ways people communicate over Instagram Direct    now is actually they share content that they find on Instagram.    So thats been a debate over and over again. But really that    decision is about keeping your feed focused on the people you    know rather than the people you know finding other stuff for    you to see. And I think that is more of a testament of our    focus on authenticity and on the connections you actually have    than about anything else.  <\/p>\n<p>    NT: So after you went to VidCon, you posted an    image on your Instagram feed of you and a bunch of celebrities  <\/p>\n<p>    KS: Totally, in fact it was a Boomerang.  <\/p>\n<p>    NT: It was a Boomerang, right! So Im going to    read some of the comments on @kevins post.  <\/p>\n<p>    KS: Sure.  <\/p>\n<p>    NT: These are the comments: Succ, Succ,    Succ me, Succ, Can you make Instagram have auto-scroll    feature? That would be awesome and expand Instagram as a app    that could grow even more, #memelivesmatter, you succ,    you can delete memes but not cancer patients, I love    #memelivesmatter, #allmemesmatter, succ, #MLM,    #memerevolution, cuck, mem, #stopthememegenocide,    #makeinstagramgreatagain, #memelivesmatter,    #memelivesmatter, mmm, gang, melon gangIm not quite    sure what all this means. Is this typical?  <\/p>\n<p>    KS: It was typical, but Id encourage you to    go to my last post which I posted for Fathers Day  <\/p>\n<p>    NT: Your last post is all nice!  <\/p>\n<p>    KS: Its all nice.  <\/p>\n<p>    NT: Theyre all about how handsome your father    is.  <\/p>\n<p>    KS: Right? Listen, he is taken. My mom is    wonderful. But there are a lot of really wonderful comments    there.  <\/p>\n<p>    NT: So why is this post from a year ago full    of cuck and #memelivesmatter and the most recent post is    full of how handsome Kevin Systroms dad is?  <\/p>\n<p>    KS: Well thats a good question. I would love    to be able to explain it, but the first thing I think is back    then there were a bunch of people who I think were unhappy    about the way Instagram was managing accounts. And there are    groups of people that like to get together and band up and    bully people, but its a good example of how someone can get    bullied, right. The good news is I run the company and I have a    thick skin and I can deal with it. But imagine youre someone    whos trying to express yourself about depression or anxiety or    body image issues and you get that. Does that make you want to    come back and post on the platform? And if youre seeing that,    does that make you want to be open about those issues as well?    No. So a year ago I think we had much more of a problem, but    the focus over that year, over both comment filtering so now    you can go in and enter your own words that basically filter    out comments that include that word. We have spam filtering    that works pretty well, so probably a bunch of those would have    been caught up in the spam filter that we have because they    were repeated comments. And also just a general awareness of    kind comments. We have this awesome campaign that we started    called #kindcomments. I dont know if you know the late night    show were they reads off mean comments on another social    platform; we started kind comments to basically set a standard    in the community that it was better and cooler to actually    leave kind comments. And now there is this amazing meme that    has spread throughout Instagram about leaving kind comments.    But you can see the marked difference between the post about    Fathers Day and that post a year ago on what technology can do    to create a kinder community. And i think were making progress    which is the important part.  <\/p>\n<p>    NT: Tell me about sort of steps one, two,    three, four, five. How do you  you dont automatically decide    to launch the seventeen things youve launched since then? Tell    me about the early conversations.  <\/p>\n<p>    KS: The early conversations were really about    what problem are we solving and we looked to the community for    stories. We talked to community members. We have a giant    community team here at Instagram, which I think is pretty    unique for technology companies. Literally, their job is to    interface with the community and get feedback and highlight    members who are doing amazing things on the platform. So    getting that type of feedback from the community about what    types of problems they were experiencing in their comments then    led us to brainstorm about all the different things we could    build. And what we realized was there was this giant wave of    machine learning and artificial intelligenceand Facebook had    developed this thing that basicallyits called deep text  <\/p>\n<p>    NT: Which launches in June of 2016, so its    right there.  <\/p>\n<p>    KS: Yup, so they have this technology and we    put two and two together and we said: You know what? I think if    we get a bunch of people to look at comments and rate them good    or badlike you go on pandora and you listen to a song, is it    good or is it badget a bunch of people to do that. Thats your    training set. And then what you do is you feed it to the    machine learning system and you let it go through 80 percent of    it and then you hold out the other 20 percent of the comments.    And then you say, Okay, machine, go and rate these comments    for us based on the training set, and then we see how well it    does and we tweak it over time, and now were at a point where    basically this machine learning can detect a bad comment or a    mean comment with amazing accuracybasically a 1 percent false    positive rate. So throughout that process of brainstorming,    looking at the technology available and then training this    filter over time with real humans who are deciding this stuff,    gathering feedback from our community and gathering feedback    from our team about how it works, were able to create    something were really proud of.  <\/p>\n<p>    NT: So when you launch it you make a very    important decision: Do you want it to be aggressive, in which    case itll probably knock out some stuff it shouldnt? Or do    you want it to be a little less aggressive, in which case a lot    of bad stuff will get through?  <\/p>\n<p>    KS: Yeah, this is the classic problem. If you    go for accuracy, you will misclassify a bunch of stuff that    actually was pretty good. So you know if your my friend and I    go on your photo and Im just joking around with you and giving    you a hard time, Instagram should let that through because    were friends and Im just giving you a hard time and thats a    funny banter back and forth. Whereas if you dont know me and I    come on and I make fun of your photo, that feels very    different. Understanding the nuance between those two is super    important and the thing we dont want to do is have any    instance where we block something that shouldnt be blocked.    The reality is its going to happen. So the question is, is    that margin of error worth it for all the really bad stuff that    gets blocked? And thats a fine balance to figure out. Thats    something were working on. We trained the filter basically to    have a one-percent false positive rate. So that means one    percent of things that get marked as bad are actually good. And    that was a top priority for us because were not here to curb    free speech, were not here to curb fun conversations between    friends, but we want to make sure we are largely attacking the    problem of bad comments on Instagram.  <\/p>\n<p>    NT: And so you go, and every comment that goes    in gets sort of run through an algorithm, and the algorithm    gives it a score from 0 to 1 on whether its likely a comment    that should be filtered or a comment that should not be    filtered, right? And then that score is combined with the    relationship of the two people?  <\/p>\n<p>    KS: No, the score actually is influenced based    on the relationship of the people  <\/p>\n<p>    NT: So the original score is influenced by,    and Instagram I believeif I have this correcthas something    like a karma score for every user, where the number of times    theyve been flagged or the number of critiques made of them is    added into something on the back end, is that goes into this    too?  <\/p>\n<p>    KS: So without getting into the magic    sauceyoure asking like Coca Cola to give up its recipeIm    going to tell you that theres a lot of complicated stuff that    goes into it. But basically it looks at the words, it looks at    our relationship, and it looks at a bunch of other signals    including account age, account history, and that kind of stuff.    And it combines all those signals and then it spits out a score    of 0 to 1 about how bad this comment is likely. And then    basically you set a threshold that optimizes for one-percent    false-positive rate.  <\/p>\n<p>    NT: when do you decide its ready to go?  <\/p>\n<p>    KS: I think at a point where the accuracy gets    to a point that internally were happy with it. So one of the    things we do here at instagram is we do this thing called    dogfoodingand not a lot of people know this term but in the    tech industry it means, you know, eat your own dog food. So    what we do is we take the products and we always apply them to    ourselves before we go out to the community. And there are    these amazing groups on Instagramand I would love to take you    through them but theyre actually all confidential but its    employees giving feedback about how they feel about specific    features.  <\/p>\n<p>    NT: So this is live on the phone to a bunch of    Instagram employees right now?  <\/p>\n<p>    KS: There are always features that are not    launched that are live on Instagram employees phones,    including things like this.  <\/p>\n<p>    NT: So theres a critique of a lot of the    advances in machine learning that the corpus on which it is    based has biases built into it. So DeepText analyzed all    Facebook commentsanalyzed some massive corpus of words that    people have typed into the internet. When you analyze those,    you get certain biases built into them. So for example, I was    reading a paper and someone had taken a corpus of text and    created a machine learning algorithm to rank restaurants, and    to look at the comments people had written under restaurants    and then to try and guess the quality of the restaurants. He    went through and he ran it, and he was like, Interesting,    because all of the Mexican restaurants were ranked badly. So    why is that? Well it turns out, as he dug deeper into the    algorithm, its because in massive corpus of text the word    Mexican is associated with illegalillegal Mexican    immigrant because that is used so frequently. And so there are    lots of slurs attached to the word Mexican, so the word    Mexican has negative connotations in the machine    learning-based corpus, which then affects the restaurant    rankings of Mexican restaurants.  <\/p>\n<p>    KS: That sounds awful  <\/p>\n<p>    NT: So how do you deal with that?  <\/p>\n<p>    KS: Well the good news is were not in the    business of ranking restaurants  <\/p>\n<p>    NT: But you are ranking sentences based on    this huge corpus of text that Facebook has analyzed as part of    DeepText  <\/p>\n<p>    KS: Its a little bit more complicated than    that. So all of our training comes from Instagram comments. So    we have hundreds of raters and its actually pretty interesting    what weve done with this set of raters: basically, human    beings that sit there and by the way human beings are not    unbiased thats not what im claimingbut you have human    beings. Each of those raters is bilingual. So they speak two    languages, they have a diverse perpsective, theyre from all    over the world. And they rank those comments basically, thumbs    up or thumbs down. Basically the instagram corpus, right?  <\/p>\n<p>    So you feed it a thumbs up, thumbs down based on an individual.    And you might say, But wait, isnt a single individual biased    in some way? Which is why we make sure every comment is    actually seen twice and given a rating twice by at least two    people to make sure that there is as minimal amount of bias in    the system as possible. And then on top of that, we also gain    feedback from not only our team but also the community, and    then were able to tweak things on the margin to make sure    things like that dont happen. Im not claiming that it wont    happenthats of course a riskbut the biggest risk of all is    doing nothing because were afraid of these things happening.    And I think its more important that we are A) aware of them,    and B) monitoring them actively, and C) making sure we have a    diverse group of raters that not only speak two languages but    are from all over the world and represent different    perspectives to make sure we have an unbiased classifier.  <\/p>\n<p>    NT: So lets take a sentence like These hos    aint loyal, which is a phrase that I believe a previous study    on Twitter had a lot of trouble with. Your theory is that some    people will say, Oh thats a lyric, therefore its okay, some    people wont know it will get through, but enough raters    looking at enough comments over time will allow lyrics to get    through, and These hoes aint loyal, I can post that on your    Instagram feed if you post a picture which deserves that    comment.  <\/p>\n<p>    KS: Well I think what I would counter is, if    you post that sentence to any person watching this, not a    single one of them would say thats a mean spirited comment to    any of us, right?  <\/p>\n<p>    NT: Right.  <\/p>\n<p>    NT: So I think thats pretty easy to get to. I    think if there are more nuance in examples, and I think thats    the spirit of your question, which is that there are grey    areas. The whole idea of machine learning is that its far    better about understanding those nuances than any algorithm has    in the past, or any single human being could. And I think what    we have to do over time is figure out how to get into that grey    area, and judge the performance of this algorithm over time to    see if it actually improves things. Because by the way, if it    causes trouble and it doesnt work, well scrap it and start    over with something new. But the whole idea here is that were    trying something. And I think a lot of the fears that youre    bringing up are warranted but is exactly why it keeps most    companies from even trying in the first place.  <\/p>\n<p>    NT: And so first youre going to launch this    filtering bad comments, and then the second thing youre going    to do is the elevation of positive comments. Tell me about how    that is going to work and why thats a priority.  <\/p>\n<p>    KS: The elevation of positive comments is more    about modeling in the system. Weve seen a bunch of times in    the system where we have this thing called the mimicry effect.    So if you raise kind comments, you actually see more kind    comments, or you see more people giving kind comments. its not    that we ever ran this test but Im sure if you raised a bunch    of mean comments you would see more mean comments. Part of this    is the piling-on effect, and I think what we can do is by    modeling what great conversations are, more people will see    Instagram as a place for that, and less for the bad stuff. And    its got this interesting psychological effect that people want    to fit in and people want to do what theyre seeing, and that    means that people are more positive over time.  <\/p>\n<p>    NT: And are you at all worried that youre    going to turn Instagram into the equivalent of an East Coast    liberal arts college?  <\/p>\n<p>    KS: I think those of us who grew up on the    East Coast might take offense to that *laughs* Im not sure    what you mean exactly.  <\/p>\n<p>    NT: I mean a place where there are trigger    warnings everywhere, where people feel like like they cant    have certain opinions, where people feel like they cant say    things. Where you put this sheen over all your conversations,    as though everything in the world is rosy and the bad stuff,    were just going to sweep it under the rug.  <\/p>\n<p>    KS: Yeah, that would be bad. Thats not    something we want. I think in the range of bad, were talking    about the lower five percent. Like the really, really, bad    stuff. I dont think were trying to play anywhere in the area    of grey. Although I realize, theres no black or white and    were going to have to play at some level. But the idea here is    to take out, I dont know, the bottom five percent of nasty    stuff. And I dont think anyone would argue that, that makes    Instagram a rosy place, it just doesnt make it a hateful    place.  <\/p>\n<p>    So you feed it a thumbs up, thumbs down based on an individual.    And you might say, But wait, isnt a single individual biased    in some way? Which is why we make sure every comment is    actually seen twice and given a rating twice by at least two    people to make sure that there is as minimal amount of bias in    the system as possible. And then on top of that, we also gain    feedback from not only our team but also the community, and    then were able to tweak things on the margin to make sure    things like that dont happen. Im not claiming that it wont    happenthats of course a riskbut the biggest risk of all is    doing nothing because were afraid of these things happening.    And I think its more important that we are A) aware of them,    and B) monitoring them actively, and C) making sure we have a    diverse group of raters that not only speak two languages but    are from all over the world and represent different    perspectives to make sure we have an unbiased classifier.  <\/p>\n<p>    NT: So lets take a sentence like These hos    aint loyal, which is a phrase that I believe a previous study    on Twitter had a lot of trouble with. Your theory is that some    people will say, Oh thats a lyric, therefore its okay, some    people wont know it will get through, but enough raters    looking at enough comments over time will allow lyrics to get    through, and These hoes aint loyal, I can post that on your    Instagram feed if you post a picture which deserves that    comment.  <\/p>\n<p>    KS: Well I think what I would counter is, if    you post that sentence to any person watching this, not a    single one of them would say thats a mean spirited comment to    any of us, right?  <\/p>\n<p>    NT: Right.  <\/p>\n<p>    NT: So I think thats pretty easy to get to. I    think if there are more nuance in examples, and I think thats    the spirit of your question, which is that there are grey    areas. The whole idea of machine learning is that its far    better about understanding those nuances than any algorithm has    in the past, or any single human being could. And I think what    we have to do over time is figure out how to get into that grey    area, and judge the performance of this algorithm over time to    see if it actually improves things. Because by the way, if it    causes trouble and it doesnt work, well scrap it and start    over with something new. But the whole idea here is that were    trying something. And I think a lot of the fears that youre    bringing up are warranted but is exactly why it keeps most    companies from even trying in the first place.  <\/p>\n<p>    NT: And you wouldnt want all of the comments    on your,You know, on your VidCon post, its a mix of sort of    jokes, and nastiness, and vapidity, and useful product    feedback. And youre getting rid of the nasty stuff, but    wouldnt it be better, if you raised like the best product    feedback and the funny jokes to the top?  <\/p>\n<p>    KS: Maybe. And maybe thats a problem well    decide to solve at some point. But right now were just focused    on making sure that people dont feel hate, you know? And I    think thats a valid thing to go after, and Im excited to do    it.  <\/p>\n<p>    NT: So the thing that interests me the most is    that its like Instagram is a world with 700 million people,    and youre writing the constitution for the world. When you get    up in the morning and you think about that power, that    responsibility, how does it affect you?  <\/p>\n<p>    KS: Doing nothing felt like the worst option    in thew world. So starting to tackle it means that we can    improve the world; we can improve the lives of as many young    people in the world that live on social media. I dont have    kids yet; I will someday, and I hope that kid, boy or girl,    grows up in a world where they feel safe online, where I as a    parent feel like theyre safe online. And you know the cheesy    saying, with great power comes great responsibility. We take on    that responsibility. And were going to go after it. But that    doesnt mean that not acting is the correct option. There are    all sorts of issues that come with acting, youve highlighted a    number of them today, but that doesnt mean we shouldnt act.    That just means we should be aware of them and we should be    monitoring them over time.  <\/p>\n<p>    NT: One of the critiques is that Instagram,    particularly for young people is very addictive. And in fact    theres a critique being made my Tristen Harris who was    a-classmate of yours, and a classmate of Mikes, and a student    in the same class as Mikes. And he says that the design of    Instagram deliberately addicts you. For example, when you open    it up it just-    KS: Sorry Im laughing just because I think the idea that    anyone inside here tries to design something that is    maliciously addictive is just so far fetched. We try to solve    problems for people and if by solving those problems for people    they like to use the product, I think weve done our job well.    This is not a casino, we are not trying to eke money out of    people in a malicious way. The idea of Instagram is that we    create something that allows them to connect with their    friends, and their family, and their interests, positive    experiences, and I think any criticism of building that system    is unfounded.  <\/p>\n<p>    NT: So all of this is aimed at making    Instagram better. And it sounds like changes so far have made    Instagram better. Is any of it aimed at making people better,    or is there any chance that the changes that happen on    Instagram will seep into the real world and maybe, just a    little bit, the conversations in this country will be more    positive than theyve been?  <\/p>\n<p>    KS: I sure hope we can stem any negativity in    the world. Im not sure we would sign up from that day one. Um,    but I actually want to challenge the initial premise which is    that this is about making Instagram better. I actually think    its about making the internet better. I hope someday the    technology that we develop and the training sets we develop and    the things we learn we can pass on to startups, we can pass on    our peers in technology, and we actually together build a    kinder, safer, more inclusive community online.  <\/p>\n<p>    NT: Will you open source the software youve    built for this?  <\/p>\n<p>    KS: Im not sure. Im not sure. I think a lot    of it comes back to how good it performs, and the willingness    of our partners to adopt it.  <\/p>\n<p>    NT: But what if this fails? What if actually    people actually get kind of turned off by instagram, they say,    Instagrams becoming like Disneyland, I dont want to be    there. And they share less?  <\/p>\n<p>    KS: The thing I love about Silicon Valley is    weve bear hugged failure. Failure is what we all start with,    we go through, and hopefully we dont end on, on our way to    success. I mean Instagram wasnt Instagram initially. It was a    failed start up before. I turned down a bunch of job offers    that would have been really awesome along the way. That was    failure. Ive had numerous product ideas at Instagram that were    totally failures. And thats okay. We bear hug it because when    you fail at least youre trying. And I think thats actually    what makes Silicon Valley different from traditional business.    Is that our tolerance for failure here is so much higher. And    thats why you see bigger risks and also bigger payoffs.  <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>Go here to read the rest:<br \/>\n<a target=\"_blank\" href=\"https:\/\/www.wired.com\/2017\/08\/instagram\/\" title=\"Instagram CEO Kevin Systrom on Free Speech, Artificial Intelligence, and Internet Addiction. - WIRED\">Instagram CEO Kevin Systrom on Free Speech, Artificial Intelligence, and Internet Addiction. - WIRED<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> Skip Article Header.  <a href=\"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/free-speech\/instagram-ceo-kevin-systrom-on-free-speech-artificial-intelligence-and-internet-addiction-wired\/\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":6,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[162384],"tags":[],"class_list":["post-212260","post","type-post","status-publish","format-standard","hentry","category-free-speech"],"_links":{"self":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/posts\/212260"}],"collection":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/users\/6"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/comments?post=212260"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/posts\/212260\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/media?parent=212260"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/categories?post=212260"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/tags?post=212260"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}