Monthly Archives: October 2015

What’s Next: Top Trends | Diary of an accidental futurist …

Posted: October 24, 2015 at 12:40 pm

Heres my re-write of the chapter on money again. I must stress that this is now going into an edit process so it will come out looking slightly different. Quite a bit of material got dumped (which really hurts when you like it). Otherwise it was a matter of better signposting, starting with the economy and then moving into money and constantly teasing out the digital. Sorry about the layout, its gone rather haywire. (BTW, Ive added a few links, which obviously wont appear in any paper versions of this chapter).

The Economy and Money: Is digital money making us more careless? The idea of the future being different from the present is so repugnant to our conventional modes of thought and behaviour that we, most of us, offer a great resistance to acting on its practice. John Maynard Keynes, economist Networks of inequality A few years ago I was walking down a street in West London when a white van glided to a halt opposite. Four men stepped out and slowly slid what looked like a giant glass coffin from the rear. Inside it was a large shark.

The sight of a live shark in London was slightly surreal, so I sauntered over to ask what was going on. It transpired that the creature in question was being installed in an underground aquarium in the basement of a house in Notting Hill. This secret subterranean lair should, I suppose, have belonged to Dr Evil. To local residents opposing deep basement developments it probably did. A more likely candidate might have been someone benefiting from the digitally networked nature of global finance. A partner at Goldman Sachs, perhaps. This is the investment bank immortalised by Rolling Stone magazine as: a great vampire squid wrapped around the face of humanity. Or possibly the owner was the trader known as the London Whale, who lost close to six billion dollars in 2012 for his employer, JP Morgan, by electronically betting on a series of highly risky and somewhat shady derivatives known as Credit Default Swaps. London real estate had become a serious place to stash funny money, so maybe the house belonged to a slippery individual dipping their fingers into the bank accounts of a corrupt foreign government or international institution. In the words of William Gibson, the former sci-fi writer, London is: Where you go if you successfully rip off your third world nation.

Whichever ruthless predator the house belonged to, something fishy was underfoot. My suspicion was that it had to do with unchecked financial liberalisation, but also how the digital revolution was turning the economy into a winner-takes-all online casino. The shift of power away from locally organised labour to globally organised capital has been occurring for a while, but the recent digital revolution has accelerated and accentuated this. Digitalisation hasnt directly enabled globalisation, but it certainly hasnt restrained it either and one of its negative side effects has been a tendency toward polarisation, both in terms of individual incomes and market monopolies.

Throughout most of modern history, around two-thirds of the money made in developed countries was typically paid as wages. The remaining third was paid as interest, dividends or other forms of rent to the owners of capital. But since 2000, the amount paid to capital has increased substantially while that paid to labour has declined, meaning that real wages have remained flat or fallen for large numbers of people.

The shift toward capital could have an innocent analogue explanation. China, home to an abundant supply of low-cost labour, has pushed wages down globally. This situation could soon reverse, as China runs out of people to move to its cities, their pool of labour shrinks, due to ageing, and Chinese wages increase. Alternatively, low-cost labour may shift somewhere else possibly Africa. But another explanation for the weakened position of human labour is that humans are no longer competing against each other, but against a range of largely unseen digital systems. It is humans that are losing out. A future challenge for governments globally will therefore be the allocation of resources (and perhaps taxes) between people and machines given that automated systems will take on an increasing number of previously human roles and responsibilities.

Same as it ever was? Ever since the invention of the wheel weve used machines to supplement our natural abilities. This has always displaced certain human skills. And for every increase in productivity and living standards thereve been downsides. Fire cooks our food and keeps us warm, but it can burn down our houses and fuel our enemys weapons. During the first industrial revolution machines further enhanced human muscle and then we outsourced further dirty and dangerous jobs to machines. More recently weve used machines to supplement our thinking by using them for tedious or repetitive tasks. Whats different now is that digital technologies, ranging from advanced robotics and sensor networks to basic forms of artificial intelligence and autonomous systems, are threatening areas where human activity or input was previously thought essential or unassailable.In particular, software and algorithms with near-zero marginal cost are now being used for higher-order cognitive tasks. This is not digital technology being used alongside humans, but as an alternative to them. This is not digital and human. This is digital instead of human. Losing an unskilled job to an expensive machine is one thing, but if highly skilled jobs are lost to cheap software where does that leave us? What skills do the majority of humans have left to sell if machines and automated systems start to think? You might be feeling pretty smug about this because you believe that your job is somehow special or terribly difficult to do, but the chances are that you are wrong, especially when you take into account whats happening to the cost and processing power of computers. Its not so much what computers are capable of now, but what they could be capable of in ten or twenty years time that you should be worried about.

I remember ten or so years ago reading that if you index the cost of robots to humans with 1990 as the base (1990=100) the cost of robots had fallen from 100 to 18.5. In contrast, the cost of people had risen to 151. Der Spiegel, a German magazine, recently reported that the cost of factory automation relative to human labour had fallen by 50 per cent since 1990. Over the shorter term there wont be much to worry about. Even over the longer term therell still be jobs that idiot savant software wont be able to do very well or do at all. But unless we wake up to the fact that were training people to compete head on with machine intelligence theres going to be trouble eventually. This is because we are filling peoples heads with knowledge thats applied according to sets of rules, which is exactly what computers do. We should be teaching people to do things that these machines cannot. We should be teaching people to constantly ask questions, find fluid problems, think creatively and act empathetically. We should be teaching high abstract reasoning, lateral thinking and interpersonal skills. If we dont a robot may one day come along with the same cognitive skills as us, but costs just $999. Thats not $999 a month, thats $999 in total. Forever. No lunch breaks, holidays, childcare, sick pay or strike action, either. How would you compete with that?

If you think thats far fetched, Foxconn, a Chinese electronics assembly company, is designing a factory in Chengdu thats totally automated no human workers involved whatsoever. Im fairly sure well eventually have factories and machines that can replicate themselves too, including software that writes its own code and 3D printers that can print other 3D-printers. Once weve invented machines that are smarter than us there is no reason to suppose that these machines wont go on to invent their own machines which we then wont be able to understand and so on ad infinitum. Lets hope these machines are nice to us. Its funny that our obsessive compulsive addiction to machines, especially mobile devices, is currently undermining our interpersonal skills and eroding our abstract reasoning and creativity, when these skills are exactly what well need to compete against the machines. But who ever said that the future couldnt be deeply ironic. There are more optimistic outcomes of course. Perhaps the productivity gains created by these new technologies will eventually show up and the resulting wealth will be more fairly shared. Perhaps therell be huge cost savings made in healthcare or education. Technologically induced productivity gains may offset ageing populations and shrinking workforces too. Its highly unlikely that humans will stop having interpersonal and social needs and even more unlikely that the delivery of all these needs will be within the reach of robots. In the shorter term its also worth recalling an insightful comment attributed to NASA in 1965, which is that: Man is the lowest-cost, all-purpose computer system which can be mass-produced by unskilled labour. But if the rewards of digitalisation are not equitable or designers decide that human agency is dispensable or unprofitable then a bleaker future may emerge, one characterised by polarisation, alienation and discomfort. Money for nothing

Tim Cook, the CEO of Apple, once responded to demands that Apple raise its return to shareholders by saying that his aim was not to make more profit. His aim was to make better products, from which greater financial returns would flow. This makes perfect sense to anyone except speculators carelessly seeking short-term financial gains at the expense of broader measures of benefit or value. As Jack Welch, the former CEO of General Electric once said: Shareholder value is the dumbest idea in the world. It was Plato who pointed out that an appetite for more could be directly linked with bad human behaviour. This led Aristotle to draw a black and white distinction between the making of things and the making of money. Both philosophers would no doubt have been disillusioned with high frequency trading algorithms algorithms being computer programs that follow certain steps to solve a problem or react to an observed situation. In 2013, algorithms traded $28 million worth of shares in 15 milliseconds after Reuters released manufacturing data milliseconds early. Doubtless money was made here, but for doing what?

Charles Handy, the contemporary philosopher, makes a similar point in his book The Second Curve that when money becomes the point of something then something goes wrong. Money is merely a secure way to hold or transmit value (or frozen desire as someone more poetically put it). Money is inherently valueless unless exchanged for something else. But the aim of many digital companies appears to be to make money by selling themselves (or their users) to someone else. Beyond this their ambition appears to be market disruption by delivering something faster or more conveniently than before. But to what end ultimately? What is their great purpose? What are they for beyond saving time and delivering customers to advertisers? In this context, high frequency trading is certainly clever, but its socially useless. It doesnt make anything other than money for small number of individuals. Moreover, while the risks to the owners of the algorithms are almost non-existent, this is not generally the case for the society as a whole. Huge profits are privatised, but huge losses tend to be socialised. Connectivity has multiple benefits, but linking things together means that any risks are linked with the result that systemic failure is a distinct possibility. So far weve been lucky. Flash Crashes such as the one that occurred on 6 May 2010 have been isolated events. On this date high-speed trading algorithms decided to sell vast amounts of stocks in seconds, causing momentary panic. Our blind faith in the power and infallibility of algorithms makes such failures more likely and more severe. As Christopher Steiner, author of Automate: How Algorithms Came to Rule Or World writes: Were already halfway towards a world where algorithms run nearly everything. As their power intensifies, wealth will concentrate towards them. Similarly, Nicholas Carr has written that: Miscalculations of risk, exacerbated by high-speed computerised trading programs, played a major role in the near meltdown of the worlds financial system in 2008. Digitalisation helped to create the sub-prime mortgage market and expanded it at a reckless rate. But negative network effects meant that the market imploded with astonishing speed, partly because financial networks were able to spread panic as easily as they had been able to transmit debt. Network effects can create communities and markets very quickly, but they can destroy them with velocity and ferocity too. Given the worlds financial markets, which influence our savings and pensions, are increasingly influenced by algorithms, this is a major cause for concern. After all, who is analysing the algorithms that are doing all of the analysing?

Out of sight and out of mind Interestingly, its been shown that individuals spend more money when they use digital or electronic money rather than physical cash. Because digital money is somehow invisible or out of sight our spending is less considered or careful. And when money belongs to someone else, a remote institution rather than a known individual for instance, any recklessness and impulsiveness is amplified. Susan Greenfield, the neuroscientist, has even gone as far as to link the 2008 financial crisis to digitalisation because digitalisation creates a mindset of disposability. If, as a trader, you have grown up playing rapid-fire computer games in digital environments you may decide that similar thrills can be achieved via trading screens without any direct real-world consequences. You can become desensitised. Looking at numbers on a screen its easy to forget that these numbers represent money and ultimately people. Having no contact with either can be consequential. Putting controls on computers can make matters worse because we tend to take less notice of information when its delivered on a screen amid a deluge of other digital distractions.

Carelessness can have other consequences too. Large basement developments such as the one I stumbled into represent more than additional living space. They are symbolic of a gap thats opening up between narcissistic individuals who believe that they can do anything they want if they can afford it and others who are attempting to hang on to some semblance of physical community. A wealthy few even take pleasure seeing how many local residents they can upset, as though it were some kind of glorious computer game. Of course, in the midst of endless downward drilling and horizontal hammering, the many have one thing that the few will never have, which is enough. Across central London, where a large house can easily cost ten million pounds, it is not unusual for basement developments to include underground car parks, gyms, swimming pools and staff quarters, although the latter are technically illegal. Its fine to stick one of natures most evolved killing creatures 50 feet underground, but local councils draw a line in the sand with Philippino nannies. The argument for downward development is centred on the primacy of the individual in modern society. Its their money (digital or otherwise) and they should be allowed to do whatever they like with it. There isnt even a need to apologise to neighbours about the extended noise, dirt and inconvenience. The argument against such developments is that its everyone elses sanity and that neighbourhoods and social cohesion rely on shared interests and some level of civility and cooperation. If people start to build private cinemas with giant digital screens in basements this means they arent frequenting public spaces such as local cinemas, which in turn impacts on the vitality of the area. In other words, an absence of reasonable restraint and humility by a handful of self-centred vulgarians limits the choices enjoyed by the broader community. This isnt totally the fault of digitalisation, far from it, but the idea that an individual can and should be left alone to do or say what they like is being amplified by digital technology. This is similar, in some respects, to the way in which being seated securely inside a car seems to bring out the worst in some drivers behaviour toward other road users. Access to technology, especially technology thats personal and mobile, facilitates remoteness, which in turn reduces the need to interact physically or consider the feelings of other human beings. Remote access, in particular, can destroy human intimacy and connection, although on the plus side such technology can be used to expose or shame individuals that do wrong in the eyes of the broader community.

In ancient Rome there was a law called Lex Sumptuaria that restrained public displays of wealth and curbed the purchasing of luxury goods. Similar sumptuary laws aimed at superficiality and excess have existed in ancient Greece, China, Japan and Britain. Perhaps its time to bring these laws back or at least to levy different rates of tax or opprobrium on immodest or socially divisive consumption or on digital products that damage the cohesiveness of the broader physical community. Income polarisation and inequality arent new. Emile Zola, the French writer, referred to rivers of money: Corrupting everyone in a fever of speculation in mid-19th Century Paris. But could it be that a fever of digital activity is similarly corrupting? Could virtualisation and personalisation be fraying the physical bonds that make us human and ultimately hold society together? Whats especially worrying here is that studies suggest that wealth beyond a certain level erodes empathy for other human beings. Perhaps the shift from physical to digital interaction and exchange is doing much the same thing. But its not just the wealthy that are withdrawing physically. Various apps are leading to what some commentators are calling the shut in economy. This is a spin-off from the on-demand economy, whereby busy people, including those that work from home, are not burdened by household chores. But perhaps hardly ever venturing outside is as damaging as physically shutting others out. As one food delivery service, Door Dash, cryptically says: Never leave home again. Where have all the jobs gone? Id like to move on to consider some other aspects of digital exchange, but before I do Id like to dig a little deeper into the question of whether computers and automated systems are creating or destroying wealth and what happens to any humans that become irrelevant to the needs of the on-demand digital economy.

The digitally networked nature of markets is making some people rich, but also spreading wealth around far more than you might think. Globally, the level of inequality between nations is lessening and so too is extreme poverty. In 1990, for example, 43 per cent of people in emerging markets lived in extreme poverty, defined as existing on less than $1 per day. By 2010, this figure had shrunk to 21 per cent. Or consider China. In 2000, around 4 per cent of Chinese households were defined as middle class. By 2012, this had increased to two-thirds and by 2022, its predicted that almost half (45%) of the Chinese population will be middle class, defined as having annual household incomes of between US $9,000 and $16,000. This has more to do with demographics and deregulation than digitalisation, but by accident or design global poverty has been reduced by half in 20 years. Nevertheless, the gap between the highest and lowest earning members of society is growing and is set to continue with the onward march of digital networks. As the novelist Jonathan Franzen says: The internet itself is in an incredibly elitist concentrator of wealth in the hands of the few while giving the appearance of voice and the appearance of democracy to people who are in fact being exploited by the technologies. If you have something that the world feels it needs right now its now possible to make an awful lot of money very quickly, especially if the need can be transmitted digitally. However, the spoils of regulatory and technological change are largely being accrued by people who are highly educated and internationally minded. If you are neither of these things then you are potentially destined for low-paid, insecure work, although at least youll have instant access to free music, movie downloads and computer games to pass the time until you die. Theres been much discussion about new jobs being invented, including jobs we cant currently comprehend, but most current jobs are fairly routine and repetitive and therefore ripe for automation. Furthermore, its unrealistic to expect that millions of people can be quickly retrained and reassigned to do jobs that are beyond the reach of robots, virtualisation and automation. Losing a few thousand jobs in car manufacturing to industrial robots or Amazon wiping out a bookshop is one thing, but what happens if automation removes vast swathes of employment across the globe? What if half of all jobs were to disappear? Moreover, if machines do most of the work how will most people acquire enough money to pay for the things that the machines make, thereby keeping the machines in employment? Maybe we should tax robots instead of people?

In theory the internet should be creating jobs. In the US between 1996 and 2005 it looked like it might. Productivity increased by around 3 per cent and unemployment fell. But by 2005 (i.e. before the global recession) this started to reverse. Why might this be so? According to McKinsey & Company, a firm of consultants, computers and related electronics, information industries and manufacturing contributed about 50 per cent of US productivity increases since 2000 but reduced (US) employment by 4,500,000 jobs. Perhaps productivity gains will take time to come. This is a common claim of techno-optimists and the authors of the book, Race Against the Machine. Looking at the first industrial revolution, especially the upheavals brought about by the great inventions of the Victorian era, they could have a point. But it could be that new technology, for all its power, cant compete with simple demographics and sovereign debt. Perhaps, for all its glitz, computing just isnt as transformative as stream power, railroads, electricity, postage stamps, the telegraph or the automobile. Yes weve got Facebook, Snapchat and Rich Cats of Instagram, but we havent set foot on the moon since 1969 and traffic in many cities moves no faster today than it did 100 years ago. It is certainly difficult to argue against certain aspects of technological change. Between 1988 and 2003, for example, the effectiveness of computers increased a staggering 43,000,000-fold. Exponentials of this nature must be creating tectonic shifts somewhere, but where exactly?

Is efficiency a good measure of value? In its heyday, in 1955, General Motors employed 600,000 people. Today, Google, a similarly iconic American company, employs around 50,000. Facebook employs about 6,000. More dramatically, when Facebook bought Instagram for $1 billion in 2012, Instagram had 30,000,000 users, but employed just 13 people full time. At the time of writing, Whatsapp had just 55 employees, but a market value exceeding that of the entire Sony Corporation. This forced Robert Reich, a former US treasury Secretary, to describe Whatsapp as: everything thats wrong with the US economy. This isnt because the company is bad its because it doesnt create jobs. Another example is Amazon. For each million dollars of revenue that Amazon makes it employs roughly one person. This is undoubtedly efficient, but is it desirable? Is it progress? These are all examples of the dematerialisation of the global economy, where we dont need as many people to produce things, especially when digital products and services have a near zero marginal cost and where customers can be co-opted as free click-workers that dont appear on any balance sheet. A handful of people are making lots of money from this and when regulatory frameworks are weak or almost non-existent and geography becomes irrelevant these sums tend to multiply. For multinational firms making money is becoming easier too, not only because markets are growing, but because huge amounts of money can be saved by using information technology to co-ordinate production and people across geographies. Technology vs. psychology If a society can be judged by how it treats those with the least then things are not looking good. Five minutes walk from the solitary shark and winner takes all mentality you can find families that havent worked in three generations. Many of them have given up hope of ever doing so. They are irrelevant to a digital economy or, more specifically, what Manuel Castells, a professor of sociology at the University of California at Berkeley, calls informational capitalism. Similarly, Japan is not far off a situation where some people will retire without ever having worked and without having moved out of the parental home. In some ways Japan is unique, for instance its resistance to immigration. But in other ways Japan offers a glimpse of what can happen when a demographic double-whammy of rapid ageing and falling fertility means that workforces shrink, pensions become unaffordable and younger generations dont enjoy the same dreams, disposable incomes or standards of living as their parents. Economic uncertainty and geo-political volatility, caused partly by a shift from analogue to digital platforms, can mean that careers are delayed, which delays marriage, which feeds through to low birth rates, which lowers GDP, which fuels more economic uncertainty. This is all deeply theoretical, but the results can be hugely human. If people dont enjoy secure employment, housing or relationships, what does this do to their physical and especially psychological state? I expect that a negative psychological shift could be the next big thing we experience unless a coherent we emerges to challenge some of the more negative aspects of not only income inequality, but the lack of secure and meaningful work for the less talented, the less skilled and less fortunate. A few decades ago people worked in a wide range of manufacturing and service industries and collected a secure salary and benefits. But now, according to Yochai Benkler, a professor at Harvard Law School, the on-demand digital economy is efficiently connecting people selling certain skills to others looking to buy. This sounds good. It sounds entrepreneurial. It sounds efficient and flexible and is perhaps an example of labour starting to develop its own capital. But its also, potentially, an example of mass consumption decoupling from large-scale employment and of the fact that unrestrained free-markets can be savagely uncompromising. Of course, unlike machines, people can vote and they can revolt too, although I think that passive disaffection and disenfranchisement are more likely. One of the great benefits of the internet has been the ease with which ideas can be transmitted across the globe, but ideas dont always turn into actions. The transmission of too much data or what might be termed too much truth is also resulting in what Castells calls: informed bewilderment. This may sound mild, but if bewilderment turns into despair and isolation theres a chance this could feed into fundamentalism, especially when the internet is so efficient at hosting communities of anger and transmitting hatred. There is also evidence emerging that enduring physical hardship and mental anguish not only create premature ageing, which compromises the immune and cardiovascular system, but that this has a lasting legacy for those people having children. This is partly because poorer individuals are more attuned to injustice, which feeds through to ill health and premature ageing, and partly because many of the subsequent diseases can be passed on genetically. But perhaps its not relative income levels per se that so offend, but the fact that its now so easy to see what you havent got. Social media spreads images of excess abundantly and exuberantly. Sites such as Instagram elicit envy and distribute depression by allowing selfies to scream: look at me! (and my oh so perfect life). Its all a lie, of course, but we dont really notice this because the internet so easily becomes a prison of belief.

A narrowing of focus In the Victorian era, when wealth was polarised, there was at least a shared moral code, broad sense of civic duty and collective responsibility. People, you might say, remained human. Nowadays, increasingly, individuals are looking out for themselves and digitalisation, true to form, is oiling the wheels of efficiency here too. Individualism has created a culture thats becoming increasingly venal, vindictive and avaricious. This isnt just true in the West. In China there is anguished discussion about individual callousness and an emergent culture of compensation. The debate was initiated back in 2011 when a toddler, Yue Yue, was hit by several vehicles in Foshan, a rapidly growing city in Guangdong province, and a video of the event was posted online. Despite being clearly hurt no vehicles stopped and nobody bothered to help until a rubbish collector picked the child up. Yue Yue later died in hospital. Another incident, also in China, saw two boys attempting to save two girls from drowning. The boys failed and were made to pay compensation of around 50,000 Yuan (about 5,000) each to the parents for not saving them. Such incidents are rare, but they are not unknown and do perhaps point toward a world that is becoming more interested in money than mankind a world that is grasping and litigious, where trust and the principle of moral reciprocity are under threat. You can argue that we are only aware of such events due to digital connectivity, which is probably true, and that both sharing and volunteering are in good health. But you can also argue that the transparency conjured up by connectivity and social media is making people more nervous about sticking their necks out. In a word with no secrets, ubiquitous monitoring and perfect remembering people have a tendency to conform. Hence we click on petitions online rather than actually doing anything. I was innocently eating my breakfast recently when I noticed that Kelloggs were in partnership with Chime for Change, an organisation committed to raise funds and awareness for girls education and empowerment through projects promoting education, health and justice. How were Kelloggs supporting this? By asking people to share a selfie to show your support. To me this is an example of internet impatience and faux familiarity. It personifies the way that the internet encourages ephemeral acts of belonging that are actually nothing of the sort. As for philanthropy, theres a lot of it around, but much of it has become, as one Museum director rather succinctly put it: money laundering for the soul. Philanthropy is becoming an offshoot of personal branding. It is buildings as giant selfies, rather than the selfless or anonymous love of humanity. One pleasing development that may offset this trend is crowd-funding, whereby individuals fund specific ideas with micro-donations. At the moment this is largely confined to inventions and the odd artistic endeavour, but theres no reason why crowds of people with small donations cant fund political or altruistic ideas or even interesting individuals with a promising future. I sometimes wonder why we havent seen a new round of revolutions in the West. Due to digital media we all know all about the haves and the have yachts. Its even easy to find out where the yachts are moored thanks to free tracking apps. Then again, we barely know our own neighbours these days, living, as we increasingly do, in digital bubbles where friends and news stories are filtered according to pre-selected criteria. The result is that we know more and more about the people and things we like, but less and less about anything, or anyone, outside of our existing preferences and prejudices. Putting aside cognitative biases such as inattentional blindness, which means we are often blissfully unaware of whats happening in front of our own eyes, theres also the thought that weve become so focussed on ourselves that focusing anger on a stranger five minutes up the road or on a distant yacht is a bit of stretch. This is especially true if you are addicted to 140 character updates of your daily existence or looking at photographs of cute cats online. Mugged by reality Is anyone out there thinking about how Marxs theory of alienation might be linked to social stratification and an erosion of humanity? I doubt it, but the fall of Communism can be connected with the dominance of individualism and the emergence of self-obsession. This is because before the fall of the Berlin Wall in 1989 there was an alternative ideology and economic system that acted as a counter-weight to the excesses of capitalism, free markets and individualism. Similarly, in many countries, an agile and attentive left took the sting out of any political right hooks. Then in the 1990s there was a dream called the internet. But the internet is fast becoming another ad-riddled venue for capitalism where, according to an early Facebook engineer (quoted by Ashlee Vance in his biography of Elon Musk) the best minds of my generation are thinking about how to make people click ads. The early dream of digital democracy has also soured because it turns out that a complete democracy of expression attracts voices that are stupid, angry and have a lot of time on their hands. This is a Jonathan Franzen again, although he reminds me of another writer, Terry Prachett, who pointed out that: real stupidity beats artificial intelligence every time. To get back to the story in hand, the point here is that if you take away any balancing forces you not only end up with tax shy billionaires, but income polarisation and casino banking. You can also end up with systemic financial crashes, another of which will undoubtedly be along shortly, thanks to our stratospheric levels of debt, the globally connected nature of risk and the corruption and villainy endemic in emerging markets. Its possible that connectivity will create calm rather than continued volatility, but I doubt it. More likely a relatively insignificant event, such as a modest rise in US interest rates, will spread panic and emotional contagion at which point anyone still living in a digital bubble will get mugged by reality. Coming back to some good news, a significant economic trend is the growth of global incomes. This sounds at odds with declining real wages, but I am talking about emerging not developed markets. According to Ernst & Young, the accountancy firm, an additional 3 billion individuals are being added to the global middle class. Thats 3 billion more smart-phone using, FitBit wearing, Linked-in profiled, Apple iCar driving, Instagram obsessives. In China, living standards have risen by an astonishing 10,000 per cent in a single generation. In terms of per capita GDP in China and India this has doubled in 16 and 12 years respectively. In the UK this took 153 years. This is pleasing, although the definition of middle class includes people earning as little as $10 a day. Many of these people also live behind the Great Firewall of China, so we shouldnt get too carried away with trickle down economics or the opening up of democracy. What globalisation giveth to jobs automation may soon be taketh away too and many may find themselves sinking downward towards working class or neo-feudal status rather than effervescently rising upward. According to Pew Research, the percentage of people in the US that think of themselves as middle class fell from 53 per cent in 2008 to 44 per cent in 2014, with 40 per cent now defining themselves as lower class compared to 25 per cent in 2008. Teachers, for example, that have studied hard, worked relentlessly and benefit society as a whole find themselves priced out of real-estate and various socio-economic classifications by the relentless rise of financial speculators. Deeper automation and virtualisation could make things worse. Martin Wolf, a Financial Times columnist, comments that intelligent machines could hollow out middle class jobs and compound inequality. Even if the newfound global wealth isnt temporary theres plenty of research to suggest that as people grow richer they focus more attention on their own needs at the expense of others. So a wealthier world may turn out to be one thats less caring. Of course, its not numbers that matter. What counts are feelings, especially feelings related to the direction of travel. The perception in the West generally is that we are mostly moving in the wrong direction. This can be seen in areas such as education and health and its not too hard to imagine a future world split into two halves, a thin, rich, well-educated, mobile elite and an overweight, poorly educated, anchored underclass. This is reminiscent of HG Wells intellectual, surface dwelling Eloi and downtrodden, subterranean Morlocks in The Time Machine and Tolkiens Mines of Moria. The only difference this time might be that its the global rich that end up living underground, cocooned from the outside world in deep basement developments. An upside to the downside Its obviously possible that this outcome is re-written. Its entirely possible that we will experience a reversal where honour, spiritual service or courage to country are valued far above commerce. This is a situation that existed in Britain and elsewhere not that long ago. Its possible that grace, humility, public spiritedness and contempt for vulgar displays of wealth could become dominant social values. Or perhaps a modest desire to leave as small a footprint as possible could become a key driving force. On the other hand, perhaps a dark dose of gloom and doom is exactly what the world needs. Perhaps the era of cheap money is coming to an end and an extended period of slow growth will do us all a world a good. A study led by Heejung Park at UCLA found that the trend towards greater materialism and reduced empathy had been partly reversed due to the 2008-2010 economic downturn. In comparison with a similar study looking at the period 20042006, US adolescents were less concerned with owning expensive items, while the importance of having a job thats worthwhile to society rose. Whether this is just cyclical or part of a permanent shift is currently impossible to say. These studies partly link with previous research suggesting that a decline in economic wealth promotes collectivism and perhaps with the idea that we only truly appreciate things when we are faced with their loss. There arent too many upsides to global pandemics, rogue asteroids and financial meltdowns, but the threat of impending death or disaster does focus the long lens of perspective, as Steve Jobs pointed out in his commencement speech at Stamford University. Digital Vs Physical trust Thats enough about the economy. How might the digitalisation of money impact our everyday behaviour in the future? I think it is still too early to make any definitive statements about particular technologies or applications, but I do believe that the extinction of cash is inevitable because digital transactions are faster and more convenient, especially for companies. Cash can be cumbersome too. Its also because governments and bureaucracies would like to reduce illegal economic activity and collect the largest amount of tax possible thereby increasing their power. In the US, for instance, its been estimated that cash costs the American economy $200 billion a year, not just due to tax evasion and theft, but also due to time wasting. A study, by Tufts University says that the average American spends 28 minutes per month traveling to ATMs, to which my reaction is so what? What are people not doing by wasting 28 minutes going to an ATM? Writing sonnets? Inventing a cure for cancer? But a wholly cashless society, or global e-currency, wont happen for a long time, partly because physical money, especially banknotes, is so tied up with notions of national identity (just look at the Euro to see how that can go wrong!). Physical money tells a rich story. It symbolises a nations heritage in a way that digital payments cannot. People in recent years have also tended to trust cash more. The physical presence of cash is deeply reassuring, especially in times of economic turmoil. In the UK, in 2012, more than half of all transactions were cash and the use of banknotes and coins rose slightly from the previous year. Why? The answer is probably that in 2012 the UK was still belt-tightening and people felt they could control their spending more easily using cash. Or perhaps people didnt trust the banks or each other. Similarly, in most rich countries, more than 90 per cent of all retail is still in physical rather than digital stores. We should also be careful not to assume that everyone is like ourselves. The people most likely to use cash are elderly, poor or vulnerable, so it would be a huge banking error, in my view, if everyone stopped accepting physical money. Its also a useful Plan B to have a stash of cash in case the economy melts down or your phone battery dies leaving you with no way to pay for dinner. This is probably swimming against the tide though and I suspect there is huge pent-up demand for mobile and automated payments. Globally cash is still king (85% of all transactions are still cash according to one recent study) but in developed economies this tends not to be the case. In the US about 60% of transactions are now digital, while in the UK non-cash payments have now overtaken physical cash. Money will clearly be made trying to get rid of physical money. According to the UK payments council the use of cash is expected to fall by a third by 2022. Nevertheless, circumstances do change and I suspect that any uptake of new payment technologies is scenario dependent. I was on the Greek island of Hydra in 2014 and much to my surprise the entire economy had reverted to physical money. This was slightly annoying, because I had just written a blog post about the death of cash based on my experience of visiting the island two years earlier. On this visit almost everywhere accepted electronic payments, but things had dramatically changed. Again, why? I initially thought the reason was Greeks attempting to avoid tax. Cash is anonymous. But it transpired that the real reason was trust. If you are a small business supplying meat to a taverna and youre worried about getting paid, you ask for cash. This is one reason why cash might endure longer than some e-evangelists tell us. Cash is a hugely convenient method to store and exchange value and has the distinct advantage of keeping our purchasing private. If we exchange physical cash for digital currency this makes it easier for companies and governments to spy on what were doing.

Countless types of cashless transactions There are many varieties of digital money. Weve had credit cards for a very long time. Transactions using cards have been digital for ages and contactless for a while. Weve grown used to private currencies, virtual currencies, micro-payments, embedded value cards, micro-payments and contactless (NFC) payments. Weve also learnt to trust PayPal and various Peer-to-Peer lending sites such as Zopa and Prosper, although one suspects that, like ATMs, we are happier taking money out than putting money in. Were also slowly getting used to the idea of payments using mobile phones. There are even a few e-exhibitionists with currency chips embedded in their own bodies and while this might take a while to catch on I can see the value in carrying around money in our bodies. A chip inserted in your jaw or arm is a bit extreme, but how about a tiny e-pill loaded with digital cash that, once swallowed, is good for $500 or about a week? Theres even digital gold, but to be honest I cant get my head around that at all. The key point here is that all of these methods of transaction are more or less unseen. They are also fast and convenient, which, I would suggest, means that spending will be more impulsive and less considered. We will have regular statements detailing our digital transactions, of course, but these will also be digital, delivered to our screens amid a deluge of other digital distractions and therefore widely ignored or not properly read.

Really thinking or mindlessly consuming? What interests me most here is whether or not attitudes and behaviours change in the presence of invisible money. There is surprisingly little research on this subject, but what does exist, along with my own experience, suggests that once we shift from physical to digital money things do change. With physical money (paper money, metal coins and cheques) we are more likely to buy into the illusion that money has inherent value. We are therefore more vigilant. In many cases, certainly my own, we are more careful. In short we think. Physical money feels real so our purchasing (and debt) is more considered. With digital money (everything from credit and debit cards to PayPal, Apple Money, iTunes vouchers, loyalty points and so forth) our spending is more impulsive. And as I said, earlier, when money is digital and belongs to someone else any careless behaviour is amplified. Quantitative Easing (QE) is perhaps a similar story. If instead of pressing a key on a computer and sending digital money to a secondary market to buy financial assets including bonds we saw fleets of trucks outside central banks being loaded with piles of real money to do the same I suspect that our reaction would be wholly different. We might even question whether a government monetising (buying) its own debt is a sensible idea given that the 2008 financial meltdown was caused by the transmission and obfuscation of debt. Of course, pumping money into assets via QE circles back to create inequality. If you own hard assets, such as real estate, then any price increases created by QE can be a good thing because it increases the value of your assets (often bought with debt, which is reduced via inflation). In contrast savers holding cash, or anyone without assets, is penalised. Its a bit of a stretch to link QE to the Arab Spring, but some people have, pointing out that food price inflation was a contributory factor, which can be indirectly linked to QEs effects on commodities. If one was a conspiracy theorist one might even suggest that QEs real aim was to drive down the value of the dollar, the pound and the Euro at the expense of spiralling hard currency debt and emerging economy currencies. Im getting back into macro-economics, which I dont want to do, but its worth pointing out that in The Downfall of Money, the author Frederick Taylor notes that Germanys hyperinflation not only destroyed the middle class, but democracy itself. As he writes, by the time inflation reached its zenith: everyone wanted a dictatorship. The cause of Germanys hyperinflation was initially Germany failing to keep up with payments due to France after WW1. But it was also caused by too much money chasing too few goods, which has shades of asset bubbles created by QE. It was depression, not inflation per se, that pushed voters toward Hitler, but this has a familiar ring. Across Europe we are seeing a significant rightwards shift and one of the main reasons why Germany wont boost the EU economy is because of the lasting trauma caused by inflation ninety years ago. If a lasting legacy of QE, debt, networked risk and a lack of financial restraint by individuals and institutions, all accentuated by digitalisation, is either high inflation or continued depression things could get nasty, in which case we might all long for the return of cash as a relatively safe and private way to endure the storm. Crypto-currency accounts The idea of a global digital economy thats free from dishonest banks, avaricious speculators and regulation-fixated governments is becoming increasingly popular, especially, as youd expect, online. Currencies around the world are still largely anchored to the idea of geographical boundaries and economies in which physical goods and services are exchanged. But what if someone invented a decentralised digital currency that operated independently of central banks? And what if that currency were to use encryption techniques, not only to ensure security and avoid confiscation or taxation, but to control the production of the currency? A crypto-currency like BitCoin perhaps?

In one scenario, BitCoin could become not only an alternative currency, but a stateless alternative payments infrastructure, competing against the like of Apple Pay and PayPal and against alternative currencies like airline miles. But theres a more radical possibility. What if a country got into trouble (Greece? Italy? Argentina?) and trust in the national currency collapsed. People might seek alternative ways to make payments or keep their money safe. If enough people flocked to something like BitCoin a government might be forced to follow suit and wed end up with a crytocurrency being used for exports, with its value tied to a particular economy or set of economies. More radically, how about a currency that rewarded certain kinds of behaviour? We have this already, in a sense, with loyalty cards, but Im thinking of something more consequential. What if the underlying infrastructure of BitCoin was used to create a currency that was distributed to people behaving in a virtuous manner? What if, for instance, money could be earned by putting more energy or water into a local network than was taken out? Or how about earning money by abstaining from the development of triple sub-basements or by visiting an elderly person that lives alone and asking them how they are? We could even pay people who smiled at strangers using eye-tracking and facial recognition technology on Apple smart glasses or Google eye-contact lenses. Given what governments would potentially be able to see and do if cash does disappear, such alternative currencies along with old-fashioned bartering could prove popular. At the moment, central banks use interest rates as the main weapon to control or stimulate the economy. But if people hoard cash because interest rates are low or because they dont trust banks then the economy is stunted. But with a cashless society the government has another weapon in its arsenal. What if banks not only charged people for holding money (negative interest rates) but governments imposed an additional levy for not spending it? This is making my head spin so we should move on to explore the brave new world of healthcare and medicine, of which money is an enabler. But before we do Id like to take a brief look at pensions and taxation and then end on considering whether the likes of Mark Zuckerberg might actually be OK really. If economic conditions are good, Id imagine that money and payments will continue to migrate toward digital formats. Alternatives to banks will spring up and governments will loosen their tax-take. However, if austerity persists, or returns, then governments will do everything they can to get hold of more of your money but they will be less inclined to spend it, especially on services. Taxation based upon income and expenditure will continue, but I expect that it will also shift towards assets and wealth and to a very real extent individual behaviour. One of the effects of moving toward digital payments and connectivity is transparency. Governments will, in theory, be able to see what youre spending your money on, but also how youre living in a broader sense. Hence stealth taxation. Have you put the wrong type of plastic in the recycling bin again? Thats a fine (tax). Kids late for school again? Fine (tax). Burger and large fries again? You get the idea Governments will seek to not only maximise revenue, but nudge people toward certain allegedly virtuous behaviours and people will be forced to pay for the tiniest transgressions. This, no doubt, will spark rage and rebellion, but theyll be a tax for that too. As for pensions, there are several plausible scenarios, but business as usual doesnt appear to be one of them. The system is a pyramid-selling scheme thats largely bust and needs to be reinvented in many countries. 1 in 7 people in the UK has no retirement savings whatsoever, for instance, and the culture of instant digital gratification would suggest that trying to get people to save a little for later wont meet with much success. What comes next largely depends on whether the culture of now persists and whether or not responsibility for the future is shared individually or collectively. If the culture of individualism and instant rewards holds firm, well end up with a very low safety net or a situation where people never fully retire. If we are able to delay gratification, well end up either with a return to a savings culture or one where the state provides significant support in return for significant contributions. The bottom line here is that pensions are set firmly in the future and while we like thinking about the future we dont like paying for it. So what might happen that could change the world for the better and make things slightly more sustainable?

An economy if people still matter In 1973, the economist EF Schumachers book Small is Beautiful warned against the dangers of gigantism. On one level the book was a pessimistic polemic about modernity in general and globalisation in particular. On the other hand it was possibly prescient and predictive. Schumacher foresaw the problem of resource constraints and foreshadowed the issue of human happiness, which he believed could not be sated by material possessions. He also argued for human satisfaction and pleasure to be central to all work, mirroring the thoughts of William Morris and the Arts and Crafts Movement. They argued that since consumer demand was such a central driver of the economy, then one way to change the world for the better would be to change what the majority of people want, which links directly back into money and our current voracious appetite for material possessions. On one level Schumachers book is still an idealistic hippy homily. On another it manages to describe our enduring desire for human scale, human relationships and technology that is appropriate, controllable and above all understandable. Physical money encourages the physical interaction of people, whereas digital cash is more hands off and remote. Digital transactions require energy and while any desire for green computing wont exactly stop the idea of a cashless society in its tracks it may yet restrain it. There are already some weak signals around this, which Schumacher may have approved of. Our desire for neo-Victorian computing (Steam Punk), craft sites like Etsy, the popularity of live music events and literature festivals and digital detoxing all point to a desire for balance and a world where humans are allowed to focus on what they do best. The partly generational shift toward temporary digital access rather than full physical ownership is also an encouraging development against what might be termed stuffocation. Schumacher also warned against the concentration of economic and political power, which he believed would lead to dehumanisation. Decisions should therefore be made on the basis of human needs rather than the revenue requirements of distantly accountable corporations and governments. In this respect the internet could go either way. It could bring people together and enable a more locally focussed and sustainable way or living or it could facilitate the growth of autocratic governments and monopolistic transnational corporations. But remember that the dematerialisation of the global economy the analogue to digital switch if you will is largely unseen and therefore mostly out of mind so very few people are discussing this at the moment. To some extent digital payments are a technology in search of a problem. Cash is easy to carry, easy to use and doesnt require a power source except to retrieve it from an ATM. Meanwhile credit and debit cards are widely accepted worldwide and online, so why do we need additional channels or formats? Maybe we dont. Maybe we dont even need money as much as we think. One of the problems with the digital economy from an economics standpoint is that digital companies dont produce many jobs. But maybe this isnt a problem. Once weve achieved shelter and security and managed to feed ourselves, the things that make us happy tend to be invisible to economists. The things that fulfil our deepest human needs are not to be physical things, but nebulous notions like love, belonging and compassion. This is reminiscent of Abraham Maslows hierarchy of needs, but unfortunately self-esteem, altruism, purpose and spirituality dont directly contribute to GDP or mass employment. Perhaps they should. It pains me to say it, but maybe the digital dreamers are onto something after all. Maybe the digital economy will change our frame of reference and focus our attention on non-monetary value and human exchange even if this goes a little crazy at times. What would Schumacher make our current economic situation? Maybe hed see the present day as the start of something nasty. Maybe hed see it as the start of something beautiful. What I suspect he would point out is that many people feel that they have lost control of their lives, especially financially. Job insecurity, austerity, debt and a lack of secure saving and pensions make people anxious. This can have physical effects. According to a study published in the medical magazine The Lancet, economic conditions can make people and their genetic dependants sick. Putting to one side the increased risk of suicide, mental health is a major casualty of volatile economic and geopolitical conditions. Psychological stress means that our bodies are flooded with stress hormones and these can make us ill. Turbulent economic conditions can also make long lasting changes to our genes, which can be a catalyst for heart disease, cancer and depression in later generations. Another study, co-led by George Slavich at the University of California at Los Angeles, says that there is historical evidence for such claims and cites the fact that generations born during recessions tend to have unusually short lifespans. Research by Jenny Tung at Duke University in North Carolina also suggests that if animals perceive they have a lower social rank, the more active their pro-inflammatory genes become. This may be applicable to humans perceiving that they are becoming digital-serfs. Even the anticipation of bad news or negative events may trigger such changes, which might explain why I recently heard that the shark in Notting Hill is now on medication.

Visit link:
What's Next: Top Trends | Diary of an accidental futurist ...

Posted in Futurist | Comments Off on What’s Next: Top Trends | Diary of an accidental futurist …

North Atlantic Treaty Organization (NATO), 1949 – 19451952 …

Posted: October 23, 2015 at 11:47 pm

North Atlantic Treaty Organization (NATO), 1949

The North Atlantic Treaty Organization was created in 1949 by the United States, Canada, and several Western European nations to provide collective security against the Soviet Union.

Signing of the NATO Treaty

NATO was the first peacetime military alliance the United States entered into outside of the Western Hemisphere. After the destruction of the Second World War, the nations of Europe struggled to rebuild their economies and ensure their security. The former required a massive influx of aid to help the war-torn landscapes re-establish industries and produce food, and the latter required assurances against a resurgent Germany or incursions from the Soviet Union. The United States viewed an economically strong, rearmed, and integrated Europe as vital to the prevention of communist expansion across the continent. As a result, Secretary of State George Marshall proposed a program of large-scale economic aid to Europe. The resulting European Recovery Program, or Marshall Plan, not only facilitated European economic integration but promoted the idea of shared interests and cooperation between the United States and Europe. Soviet refusal either to participate in the Marshall Plan or to allow its satellite states in Eastern Europe to accept the economic assistance helped to reinforce the growing division between east and west in Europe.

In 19471948, a series of events caused the nations of Western Europe to become concerned about their physical and political security and the United States to become more closely involved with European affairs. The ongoing civil war in Greece, along with tensions in Turkey, led President Harry S. Truman to assert that the United States would provide economic and military aid to both countries, as well as to any other nation struggling against an attempt at subjugation. A Soviet-sponsored coup in Czechoslovakia resulted in a communist government coming to power on the borders of Germany. Attention also focused on elections in Italy as the communist party had made significant gains among Italian voters. Furthermore, events in Germany also caused concern. The occupation and governance of Germany after the war had long been disputed, and in mid-1948, Soviet premier Joseph Stalin chose to test Western resolve by implementing a blockade against West Berlin, which was then under joint U.S., British, and French control but surrounded by Soviet-controlled East Germany. This Berlin Crisis brought the United States and the Soviet Union to the brink of conflict, although a massive airlift to resupply the city for the duration of the blockade helped to prevent an outright confrontation. These events caused U.S. officials to grow increasingly wary of the possibility that the countries of Western Europe might deal with their security concerns by negotiating with the Soviets. To counter this possible turn of events, the Truman Administration considered the possibility of forming a European-American alliance that would commit the United States to bolstering the security of Western Europe.

Signing of the Brussels Treaty

The Western European countries were willing to consider a collective security solution. In response to increasing tensions and security concerns, representatives of several countries of Western Europe gathered together to create a military alliance. Great Britain, France, Belgium, the Netherlands and Luxembourg signed the Brussels Treaty in March, 1948. Their treaty provided collective defense; if any one of these nations was attacked, the others were bound to help defend it. At the same time, the Truman Administration instituted a peacetime draft, increased military spending, and called upon the historically isolationist Republican Congress to consider a military alliance with Europe. In May of 1948, Republican Senator Arthur H. Vandenburg proposed a resolution suggesting that the President seek a security treaty with Western Europe that would adhere to the United Nations charter but exist outside of the Security Council where the Soviet Union held veto power. The Vandenburg Resolution passed, and negotiations began for the North Atlantic Treaty.

In spite of general agreement on the concept behind the treaty, it took several months to work out the exact terms. The U.S. Congress had embraced the pursuit of the international alliance, but it remained concerned about the wording of the treaty. The nations of Western Europe wanted assurances that the United States would intervene automatically in the event of an attack, but under the U.S. Constitution the power to declare war rested with Congress. Negotiations worked toward finding language that would reassure the European states but not obligate the United States to act in a way that violated its own laws. Additionally, European contributions to collective security would require large-scale military assistance from the United States to help rebuild Western Europes defense capabilities. While the European nations argued for individual grants and aid, the United States wanted to make aid conditional on regional coordination. A third issue was the question of scope. The Brussels Treaty signatories preferred that membership in the alliance be restricted to the members of that treaty plus the United States. The U.S. negotiators felt there was more to be gained from enlarging the new treaty to include the countries of the North Atlantic, including Canada, Iceland, Denmark, Norway, Ireland, and Portugal. Together, these countries held territory that formed a bridge between the opposite shores of the Atlantic Ocean, which would facilitate military action if it became necessary.

President Truman inspecting a tank produced under the Mutual Defense Assistance Program

The result of these extensive negotiations was the signing of the North Atlantic Treaty in 1949. In this agreement, the United States, Canada, Belgium, Denmark, France, Iceland, Italy, Luxemburg, the Netherlands, Norway, Portugal, and the United Kingdom agreed to consider attack against one an attack against all, along with consultations about threats and defense matters. This collective defense arrangement only formally applied to attacks against the signatories that occurred in Europe or North America; it did not include conflicts in colonial territories. After the treaty was signed, a number of the signatories made requests to the United States for military aid. Later in 1949, President Truman proposed a military assistance program, and the Mutual Defense Assistance Program passed the U.S. Congress in October, appropriating some $1.4 billion dollars for the purpose of building Western European defenses.

Soon after the creation of the North Atlantic Treaty Organization, the outbreak of the Korean War led the members to move quickly to integrate and coordinate their defense forces through a centralized headquarters. The North Korean attack on South Korea was widely viewed at the time to be an example of communist aggression directed by Moscow, so the United States bolstered its troop commitments to Europe to provide assurances against Soviet aggression on the European continent. In 1952, the members agreed to admit Greece and Turkey to NATO and added the Federal Republic of Germany in 1955. West German entry led the Soviet Union to retaliate with its own regional alliance, which took the form of the Warsaw Treaty Organization and included the Soviet satellite states of Eastern Europe as members.

The collective defense arrangements in NATO served to place the whole of Western Europe under the American nuclear umbrella. In the 1950s, one of the first military doctrines of NATO emerged in the form of massive retaliation, or the idea that if any member was attacked, the United States would respond with a large-scale nuclear attack. The threat of this form of response was meant to serve as a deterrent against Soviet aggression on the continent. Although formed in response to the exigencies of the developing Cold War, NATO has lasted beyond the end of that conflict, with membership even expanding to include some former Soviet states. It remains the largest peacetime military alliance in the world.

Read more here:
North Atlantic Treaty Organization (NATO), 1949 - 19451952 ...

Posted in NATO | Comments Off on North Atlantic Treaty Organization (NATO), 1949 – 19451952 …

Gravitational singularity – Wikipedia, the free encyclopedia

Posted: at 9:44 am

A gravitational singularity or spacetime singularity is a location where the force of gravity has become effectively infinite, and the quantities that are used to measure the gravitational field of a singularity become infinite in a way that does not depend on the coordinate system from the standpoint of any observer of the singularity. These quantities are the scalar invariant curvatures of spacetime, which includes a measure of the density of matter. The laws of normal spacetime could not exist within a singularity and it is currently postulated that matter cannot cross the event horizon of a singularity due to the effects of time dialation.[1][2][3] Singularities are theorized to exist at the center of Black Holes, within Cosmic Strings, and as leftover remants from the early formation of the Universe following the Big Bang. Although gravitational singularities were proposed by Einstein in his General Relativity Theory, their existence has not been confirmed. [4][5][6][7]

For the purposes of proving the PenroseHawking singularity theorems, a spacetime with a singularity is defined to be one that contains geodesics that cannot be extended in a smooth manner.[8] The end of such a geodesic is considered to be the singularity. This is a different definition, useful for proving theorems.[9][10]

The two most important types of spacetime singularities are curvature singularities and conical singularities.[11] Singularities can also be divided according to whether or not they are covered by an event horizon (naked singularities are not covered).[12] According to modern general relativity, the initial state of the universe, at the beginning of the Big Bang, was a singularity.[13] Both general relativity and quantum mechanics break down in describing the Big Bang,[14] but in general, quantum mechanics does not permit particles to inhabit a space smaller than their wavelengths (See: Wave-particle_duality).[15]

Another type of singularity predicted by general relativity is inside a black hole: any star collapsing beyond a certain point (the Schwarzschild radius) would form a black hole, inside which a singularity (covered by an event horizon) would be formed, as all the matter would flow into a certain point (or a circular line, if the black hole is rotating).[16] This is again according to general relativity without quantum mechanics, which forbids wavelike particles entering a space smaller than their wavelength. These hypothetical singularities are also known as curvature singularities.

In theoretical modeling with supersymmetry theory, a singularity in the moduli space (a geometric space using coordinates to model objects, observers, or locations) happens usually when there are additional massless degrees of freedom (dimensions) at a certain point. Similarly, in String Theory and in M Theory, it is thought that singularities in spacetime often mean that there are additional degrees of freedom (physical dimensions beyond the four dimensions described by General Relativity) that exist only within the vicinity of the singularity. The same fields related to the whole spacetime are postulated to also exist according to this theory; for example, the electromagnetic field. In known examples of string theory, the latter degrees of freedom are related to closed strings, while the degrees of freedom are "stuck" to the singularity and related either to open strings or to the twisted sector of an orbifold (A theoretical construct of abstract mathematics). This is however, only a theory.[17][18]

A theory supported by Stephen Hawkings called the Black hole information paradox postulates that matter cannot cross the event horizon of a singularity or black hole and remains as stored information just beyond the event horizon and is slowly released as Hawking radiation or held at the event horizon permanently due to the effects of time dialation. "The information is not stored in the interior of the black hole as one might expect, but in its boundary - the event horizon," he told a conference at the KTH Royal Institute of Technology in Stockholm, Sweden. (meaning that as matter enters the event horizon of a black hole the deeper it travels inside the black hole, the slower time flows for that matter relative to an observer outside the black hole watching the matter travel through the event horizon. Time essentially slows until it virtually stops as the matter reaches the event horizon of the singularity and can never make it to the center and is held there forever).[19]

Solutions to the equations of general relativity or another theory of gravity (such as supergravity) often result in encountering points where the metric blows up to infinity. However, many of these points are completely regular, and the infinities are merely a result of using an inappropriate coordinate system at this point (meaning that Einsteins partial differential equations that describe spacetime curvature and gravity produce infinite values if you provide bad data). In order to test whether there is a singularity at a certain point, one must check whether at this point diffeomorphism invariant quantities (coordinates in a coordinate system describing an observer in such a way that the relationships of the physical law being tested do not vary based on the coordinates of the observer being at different locations) which are scalars become infinite (a scalar is a pure number representing a value, like length). Such quantities are the same in every coordinate system, so these infinities will not "go away" by a change of coordinates. (when proper coordinate systems are used to describe an observers location, no matter what system is employed, a singularity will always produce these infinites using Einstein's partial differential equations to describe space time curvature).

An example is the Schwarzschild solution that describes a non-rotating, uncharged black hole. In coordinate systems convenient for working in regions far away from the black hole, a part of the metric becomes infinite at the event horizon. However, spacetime at the event horizon is regular. The regularity becomes evident when changing to another coordinate system (such as the Kruskal coordinates), where the metric is perfectly smooth. On the other hand, in the center of the black hole, where the metric becomes infinite as well, the solutions suggest a singularity exists. The existence of the singularity can be verified by noting that the Kretschmann scalar, being the square of the Riemann tensor i.e. , which is diffeomorphism invariant, is infinite. While in a non-rotating black hole the singularity occurs at a single point in the model coordinates, called a "point singularity". In a rotating black hole, also known as a Kerr black hole, the singularity occurs on a ring (a circular line), known as a "ring singularity". Such a singularity may also theoretically become a wormhole.[20]

A conical singularity occurs when there is a point where the limit of every diffeomorphism invariant quantity is finite, in which case spacetime is not smooth at the point of the limit itself. Thus, spacetime looks like a cone around this point, where the singularity is located at the tip of the cone. The metric can be finite everywhere if a suitable coordinate system is used. An example of such a conical singularity is a cosmic string. Cosmic strings are theoretical, and their existence has not yet been confirmed. [21]

Until the early 1990s, it was widely believed that general relativity hides every singularity behind an event horizon, making naked singularities impossible. This is referred to as the cosmic censorship hypothesis. However, in 1991, physicists Stuart Shapiro and Saul Teukolsky performed computer simulations of a rotating plane of dust that indicated that general relativity might allow for "naked" singularities. What these objects would actually look like in such a model is unknown. Nor is it known whether singularities would still arise if the simplifying assumptions used to make the simulation were removed.[22][23][24]

Some theories, such as the theory of loop quantum gravity suggest that singularities may not exist. The idea is that due to quantum gravity effects, there is a minimum distance beyond which the force of gravity no longer continues to increase as the distance between the masses becomes shorter.[25][26]

The Einstein-Cartan-Sciama-Kibble theory of gravity naturally averts the gravitational singularity at the Big Bang. This theory extends general relativity to matter with intrinsic angular momentum (spin) by removing a constraint of the symmetry of the affine connection and regarding its antisymmetric part, the torsion tensor, as a variable in varying the action. The minimal coupling between torsion and Dirac spinors generates a spinspin interaction in fermionic matter, which becomes dominant at extremely high densities and prevents the scale factor of the Universe from reaching zero. The Big Bang is replaced by a cusp-like Big Bounce at which the matter has an enormous but finite density and before which the Universe was contracting (what is theorized is that matter exerts a counterforce based upon the spin (angular momentum) which is present in all fermionic matter that will resist the effects of gravity beyond a certain point of compression and a singularity can never fully form). [27]

Read the rest here:
Gravitational singularity - Wikipedia, the free encyclopedia

Posted in The Singularity | Comments Off on Gravitational singularity – Wikipedia, the free encyclopedia

Origins of Eugenics: From Sir Francis Galton to Virginias …

Posted: at 9:44 am

Sir Francis Galton. Courtesy of the American Philosophical Society. [2.1]

ENLARGE [2.2] Faces and Races, illustration from a eugenical text, Racial History of Mankind. Courtesy of Special Collections, Pickler Memorial Library, Truman State University.

[2.3] Harry H. Laughlin and Charles Davenport at the Eugenics Record Office. Courtesy of Cold Spring Harbor Laboratory Archives.

Sir Francis Galton first coined the term eugenics in 1883. Put simply, eugenics means well-born. Initially Galton focused on positive eugenics, encouraging healthy, capable people of above-average intelligence to bear more children, with the idea of building an improved human race. Some followers of Galton combined his emphasis on ancestral traits with Gregor Mendels research on patterns of inheritance, in an attempt to explain the generational transmission of genetic traits in human beings.

Negative eugenics, as developed in the United States and Germany, played on fears of race degeneration. At a time when the working-class poor were reproducing at a greater rate than successful middle- and upper-class members of society, these ideas garnered considerable interest. One of the most famous proponents in the United States was President Theodore Roosevelt, who warned that the failure of couples of Anglo-Saxon heritage to produce large families would lead to race suicide.

The center of the eugenics movement in the United States was the Eugenics Record Office (ERO) at Cold Spring Harbor, New York. Biologist Charles Davenport established the ERO, and was joined in his work by Director Harry H. Laughlin. Both men were members of the American Breeders Association. Their view of eugenics, as applied to human populations, drew from the agricultural model of breeding the strongest and most capable members of a species while making certain that the weakest members do not reproduce.

Eugenicists attempted to demonstrate the power of heredity by constructing pedigree charts of defective families. These charts were used to scientifically quantify the assertion that human frailties such as profligacy and indolence were genetic components that could be passed from one generation to the next. Two studies were published that charted the propensity towards criminality, disease, and immoral behavior of the extended families of the Jukes and the Kallikaks. Eugenicists pointed to these texts to demonstrate that feeblemindedness was an inherited attribute and to reveal how the care of such degenerates represented an enormous cost to society.

The ERO promoted eugenics research by compiling records or pedigrees of thousands of families. Charles Davenport created The Family History Book, which assisted field workers as they interviewed families and assembled pedigrees specifying inheritable family attributes which might range from allergies to civic leadership. Even a propensity for carpentry or dress-making was considered a genetically inherited trait. Davenport and Laughlin also issued another manual titled How to Make a Eugenical Family Study to instruct field workers in the creation of pedigree charts of study subjects from poor, rural areas or from institutionalized settings. Field workers used symbols to depict defective conditions such as epilepsy and sexual immorality.

The American Eugenics Society presented eugenics exhibits at state fairs throughout the country, and provided information encouraging high-grade people to reproduce at a greater rate for the benefit of society. The Society even sponsored Fitter Family contests.

ENLARGE [2.4] Kallikak family of New Jersey Normal and Degenerate Lines (enlarge to view additional eugenical pedigree charts). Courtesy of Paul Lombardo.

ENLARGE [2.5] Eugenics Display. Courtesy of the American Philosophical Society.

[2.6] Winners of Fittest Family Contest. Courtesy of the American Philosophical Society.

[2.7] Harry H. Laughlin photograph. Courtesy of American Philosophical Society.

ENLARGE [2.8] Comparative Intelligence Chart. Courtesy of the American Philosophical Society.

ENLARGE [2.9] Virginias Racial Integrity Act of 1924 (enlarge to view additional Virginia legislative acts). Courtesy of Special Collections, Pickler Memorial Library, Truman State University.

In 1914, Harry H. Laughlin attended the first Race Betterment Conference, sponsored by J. H. Kellogg. The same year, in his Model Sterilization Law, Laughlin declared that the socially inadequate of society should be sterilized. This Model Law was accompanied by pedigree charts, which were used to demonstrate the hereditary nature of traits such as alcoholism, illegitimacy, and feeblemindedness. Laughlin asserted that passage of these undesirable traits to future generations would be eradicated if the unfortunate people who possessed them could be prevented from reproducing. In 1922 Laughlins Model Law was included in the book Eugenical Sterilization in the United States. This book compiled legal materials and statistics regarding sterilization, and was a valuable reference for sterilization activists in states throughout the country.

Proponents of eugenics worked tirelessly to assert the legitimacy of this new discipline. For Americans who feared the potential degradation of their race and culture, eugenics offered a convenient and scientifically plausible response to those fears. Sterilization of the unfit seemed a cost-effective means of strengthening and improving American society.

By 1924 Laughlins influence extended in several directions. He testified before Congress in support of the Immigration Restriction Act to limit immigration from eastern and southern Europe. Laughlin influenced passage of this law by presenting skewed data to support his assertion that the percentage of these immigrant populations in prisons and mental institutions was far greater than their percentage in the general population would warrant.

Laughlin also provided guidance in support of Virginias Racial Integrity Act, which made it illegal for whites in Virginia to marry outside their race. The act narrowly defined who could claim to be a member of the white race stating that the term white person shall apply only to such person as has no trace whatever of any blood other than Caucasian. Virginia lawmakers were careful to leave an escape clause for colleagues who claimed descent from Pocahontasthose with 1/16 or less of the blood of the American Indian would also count as white.

The language of Laughlins Model Sterilization Act was used in Virginias Eugenical Sterilization Act to legalize compulsory sterilizations in the state. This legislation to rid Virginia of defective persons was drafted by Aubrey E. Strode, a former member of the Virginia General Assembly, at the request of longtime associate, Albert Priddy, who directed the Virginia Colony for the Epileptic and Feebleminded in Lynchburg, Virginia.

2004 Claude Moore Health Sciences Library

Excerpt from:
Origins of Eugenics: From Sir Francis Galton to Virginias ...

Posted in Eugenics | Comments Off on Origins of Eugenics: From Sir Francis Galton to Virginias …

Upstaged NATO searches for ‘360-degree’ response to Russia

Posted: at 9:43 am

TRAPANI, Italy The brass band played, the flags waved and Western generals delivered speeches brimming with resolve as NATO began big war games in the central Mediterranean this week.

But the military display seemed faintly unreal while Russian warplanes were bombing Syrian rebels a few hundred miles to the east in a coordinated action with President Bashar al-Assad's armed forces and Iran's Revolutionary Guards.

NATO, which waged an air campaign to help Libyan rebels oust Muammar Gaddafi, then left that country to descend into anarchy, is not a player in Syria and is watching uncomfortably as its former Cold War adversary Russia widens its role there.

The speed and scope of Moscow's intervention in Syria's four-year-old civil war, coming after Russia's seizure of Crimea and support for pro-Kremlin rebels in eastern Ukraine last year, wrong-footed the U.S.-led alliance and has heightened soul-searching about its future.

"The West has been tactically surprised. I don't think they anticipated what (Russian President Vladimir) Putin would get up to," said Nick Witney, a former European Defence Agency chief now at the European Council on Foreign Relations.

NATO last year set in motion its biggest modernization since the Cold War. But the alliance's political and military elite now see the need for a broader plan that goes beyond deterring Russia in the east. They call it thinking "360 degrees".

"We need to develop a strategy for all kinds of crises, at 360 degrees," said Gen. Denis Mercier, the Frenchman who heads NATO's command focused on future threats. "We need to react in the south, in the east, the north, all around."

NATO's problem is that such a strategy is still embryonic while developments in Europe's neighborhood are moving faster than the ponderous approach of the 28-nation defense pact, created in 1949 to deter the Soviet threat.

From the Baltics, where Russia has a naval base in Kaliningrad, through the Black Sea and annexed Crimea, to Syria, Moscow has stationed anti-aircraft and anti-ship missiles able to cover huge areas.

NATO officials see the emergence of a strategy of defensive zones of influence, with surface-to-air missile batteries and anti-ship missiles that could disrupt NATO moving across air, land and sea or deny it access to some areas.

Unconventional warfare techniques are part of the equation, ranging from unidentified troops - the so-called "green men" without insignia on their uniforms seen in Crimea and eastern Ukraine - to disinformation operations and cyber attacks.

OVERESTIMATING RUSSIA?

NATO also faces failing states, war, Islamist militancy and a refugee crisis at Europe's borders. That is partly a result of the European Union's inability to stabilize its neighborhood economically.

But critics say it is also due to U.S. President Barack Obama's aversion to entanglement in Middle East wars in the wake of the U.S. invasion of Iraq in 2003. That has led to a decline in Washington's influence across the region.

While NATO is drawing up a multi-layered deterrence plan, officials acknowledge a risk that Russia might again move faster to pre-empt Western action. For instance, it could move warships from the Eastern Mediterranean to the Libyan coast to hamper any possible NATO effort to support a government of national unity in the future.

Still, some say NATO has been here before and any talk of a lack of preparedness is overblown. Past bouts of questioning of the alliance's relevance led to operations in the Balkans and in Afghanistan - a significant departure from 40 years of Cold War deterrence in which NATO forces never operated "out of area".

A NATO official rejected any suggestion the alliance was passively watching Russia's military build-up in Syria, noting that three allies the United States, France and Turkey were involved individually in the coalition waging air strikes against Islamic State rebels in Syria and Iraq.

Some experts see a danger of overestimating Putin, who oversees an economy weakened by Western sanctions and lower oil prices and cannot match NATO military power over the long term.

"We should be under no illusions about Putin's hostility to the West but also be very careful not to over react to what a damaged Russian economy can produce in the way of military capability," said Witney, a former British defense planner.

LOOKING TO AMAZON, DHL

NATO's public response is to test its new spearhead force of 5,000 troops, ready to move within a few days. Over the next five weeks, the alliance is carrying out its biggest military exercises since 2002, with 36,000 troops, 230 military units, 140 aircraft and more than 60 ships, to certify the force.

Such measures, agreed at a NATO summit in Wales last year following Russia's annexation of Crimea, are aimed chiefly at reassuring eastern allies that Russia will not be able to invade them too. There is still debate about whether the spearhead force could be used in North Africa or beyond.

Small command posts with NATO flags from Estonia to Bulgaria and the spearhead force are ready. But one NATO diplomat called such measures "the minimum necessary", and Gen. Mercier said the so-called Readiness Action Plan was "just a first step".

"We have worked on reassuring our allies," said Gen. Philip Breedlove, NATO's supreme commander in Europe. "We are not exactly sure what it will take to work in the future," he said when asked what NATO's modern deterrents might look like.

The next NATO summit next July in Warsaw is the target date for proposals for more modern, agile deterrents.

Such ideas include setting up NATO-flagged command posts on the southern flank and adapting the spearhead force for maritime and air operations.

They could also feature a permanent naval force to patrol the Mediterranean and work more closely with the European Union and the United Nations in stabilizing fragile states.

Another idea involves including a nuclear deterrent in training exercises, something Britain supports but others, such as Germany, worry would be seen as a provocation by Russia.

NATO's Gen. Mercier even suggested looking to companies such as courier DHL Worldwide Express and online retailer Amazon.com to improve NATO's deployment speed.

"The question is how to have new ideas to make deployments easier. We should look at what the civilian world does, to DHL and Amazon. How do they improve their logistics?" Mercier said.

(Reporting by Robin Emmott; Editing by Paul Taylor)

Read the original:
Upstaged NATO searches for '360-degree' response to Russia

Posted in NATO | Comments Off on Upstaged NATO searches for ‘360-degree’ response to Russia

NATO: Pictures, Videos, Breaking News

Posted: at 12:49 am

Marina Kaljurand is Estonia's Minister of Foreign Affairs. Prior to becoming Foreign Minister in July 2015, she was a long-time Estonian diplomat, and an ambassador to the United States, Mexico, Canada, Russia, Kazakhstan and Israel.

From Syria to Ukraine, a pesky and newly reinvigorated Vladimir Putin is testing the west. In this new geopolitical face-off, Poland has assumed key strategic importance and will no doubt play a hugely significant diplomatic and even military role.

Nikolas Kozloff

Author, 'Revolution! South America and the Rise of the New Left'

In the hell of bad solutions on offer for Syria, some are worse than others. And the one devised by Russian President Vladimir Putin is probably the most infernal of all.

Recent developments confirm a major shift in the balance of power in the Middle East. It appears to me that Russia is dictating the pace of events, raising the question of whether Syria is becoming a proxy war between the United States and Russia.

Since Russian President Vladamir Putin began his military air campaign in Syria last week, Western media has been clamoring to explain his motivation. Most of the analysis so far has been incomplete or half-baked at best.

No one should have been surprised that Russia committed their military to the task of saving their ally in Syria from defeat. And no one should now be surprised if Saudi Arabia steps up its support for the Syrian opposition; or if the opposition attacks Russian forces in Syria.

James Zogby

President, Arab American Institute; author, 'Arab Voices'

The challenge of Putin as well as ISIS requires an answer beyond avoidance and containment. The threat is immediate but also the challenge to the rule of law and the ideology upon which free and democratic states have prospered as societies and economies over the last few decades.

Information warfare requires an infrastructure of broadcasting, social media, and other communication assets that can direct messages to the same audiences the Russians target, but do so more effectively.

Philip Seib

Professor, University of Southern California

LONDON -- The refugee crisis that dominates Europe's TVs and newspapers is the product of the horrendous civil war that still rages in Syria. Why will we not focus our attention on this? The reality in Syria is that the war creates the refugees. Do more to stop this war.

No doubt, the bombastic Donald is an unlikely president. Yet what may be most extraordinary about his campaign is that on foreign policy, at least, he may be the most sensible Republican in the race.

Turkey's offering Washington a fig leaf of cooperation against the Islamic State, but it's turning all its firepower against the most effective anti-ISIS fighters in the region -- the Kurds.

The challenge for the European Union and its member states, particularly Germany, is in balancing the often incongruous demands of co-operation and self-interest, and thus demonstrate to their own citizens that concrete achievements can still create a Europe of solidarity and prosperity as Schuman envisaged.

David Miles

Managing Editor of Global Politics, Carnegie Scholar, Associate Fellow at The Centre for Global Constitutionalism

As a Jew, Schumer is not allowed to "break ranks" and make up his own mind based on his clear thinking. In doing so, he is clearly an "Israel-firster" and a "Netanyahu marionette".

Shahar Azani

Executive Director of StandWithUs Northeast U.S, a Global Israel Education Organization. Former diplomat. Loves Israel. In love with Life.

The headlines on the talks between Brussels and Athens on the 86 billion euro bailout of Greece seem to indicate nothing but problems. It appears that even before talks can begin, Greece is being asked to enact further laws.

It was seventy years ago that human beings were first targeted with -- and annihilated by -- an atomic bomb.

The Turkish President's self-serving fake war against terrorism could have the tragic consequence of escalating the violence throughout Turkey and neighboring countries. If Ankara is truly interested in countering the Jihadists, it should have done that long ago, instead of arming and abetting ISIS and other terror groups.

Link:
NATO: Pictures, Videos, Breaking News

Posted in NATO | Comments Off on NATO: Pictures, Videos, Breaking News

Genome Journal – NRC Research Press

Posted: at 12:47 am

This site uses cookies to improve performance. If your browser does not accept cookies, you cannot view this site.

There are many reasons why a cookie could not be set correctly. Below are the most common reasons:

This site uses cookies to improve performance by remembering that you are logged in when you go from page to page. To provide access without cookies would require the site to create a new session for every page you visit, which slows the system down to an unacceptable level.

This site stores nothing other than an automatically generated session ID in the cookie; no other information is captured.

In general, only the information that you provide, or the choices you make while visiting a web site, can be stored in a cookie. For example, the site cannot determine your email name unless you choose to type it. Allowing a website to create a cookie does not give that or any other site access to the rest of your computer, and only the site that created the cookie can read it.

Go here to read the rest:
Genome Journal - NRC Research Press

Posted in Genome | Comments Off on Genome Journal – NRC Research Press

What’s a Genome? – Genome News Network

Posted: at 12:47 am

A genome is all of a living thing's genetic material. It is the entire set of hereditary instructions for building, running, and maintaining an organism, and passing life on to the next generation. The whole shebang.

In most living things, the genome is made of a chemical called DNA. The genome contains genes, which are packaged in chromosomes and affect specific characteristics of the organism.

Imagine these relationships as a set of Chinese boxes nested one inside the other. The largest box represents the genome. Inside it, a smaller box represents the chromosomes. Inside that is a box representing genes, and inside that, finally, is the smallest box, the DNA.

In short, the genome is divided into chromosomes, chromosomes contain genes, and genes are made of DNA.

The word " genome " was coined in about 1930, even though scientists didn't know then what the genome was made of. They only knew that the genome was important enough, whatever it was, to have a name.

Each one of earth's species has its own distinctive genome: the dog genome, the wheat genome, the genomes of the cow, cold virus, bok choy, Escherichia coli (a bacterium that lives in the human gut and in animal intestines), and so on.

So genomes belong to species, but they also belong to individuals. Every giraffe on the African savanna has a unique genome, as does every elephant, acacia tree, and ostrich. Unless you are an identical twin, your genome is different from that of every other person on earthin fact, it is different from that of every other person who has ever lived.

Though unique, your genome is still recognizably a human genome. The difference is simply a matter of degree: The genome differences between two people are much smaller than the genome differences between people and our closest relatives, the chimpanzees.

. . . . . . . . . . . . . . . . . . . . .

Updated on January 15, 2003

See the article here:
What's a Genome? - Genome News Network

Posted in Genome | Comments Off on What’s a Genome? – Genome News Network

Genome dictionary definition | genome defined

Posted: at 12:47 am

noun

Genome is defined as all of a somatic cell's genetic information, or a set of haploid chromosomes.

An example of a genome is what determines the physical characteristics of a person.

Origin of genome

also genom

noun

Origin of genome

Related Forms:

(plural genomes)

From German Genom; gene + -ome

SentencesSentence examples

But this (a listing of illustrations of how the genome can change other than by mutations) doesnt constitute a crisis its a very interesting finding that shows that variation in a genome can arise by processes other than mutation of an organisms own DNA. The disposition of that variation still must occur via either natural selection (it can be good or bad) or genetic drift (no effect on fitness). This hasnt really changed the theory of evolution one iota, though its changed our view of where organisms can acquire new genes. Jerry Coyne

Continued here:
Genome dictionary definition | genome defined

Posted in Genome | Comments Off on Genome dictionary definition | genome defined

Genome | Broad Institute of MIT and Harvard

Posted: at 12:47 am

A genome is the full set of instructions needed to make every cell, tissue, and organ in your body. Almost every one of your cells contains a complete copy of these instructions, written in the four-letter language of DNA (A, C, T, and G). The human genome contains 3 billion of these "letters" or bases. This means that if your genome were written out on sheets of paper and stacked as books, the tower of tomes would be almost as high as the Washington Monument!

If you think of the human genome as an encyclopedia, the information it contains is divided into 23 volumes, called chromosomes. Each chromosome contains genes - "sentences" of genetic instructions that tell the cell how to make proteins. We know the human genome contains about 20,500 of these genes, but the meaning of much of the remaining text within it is a mystery.

Surprisingly, the human genome is not static. Throughout life, exposure to certain substances - such as X-rays, sunlight, chemicals, and more - can begin to subtly change the genome in some cells. If a cell acquires a set of genomic changes that allows it to grow out of control, invade surrounding tissue, and spread to other sites in the body, cancer develops. A cancer patient is thought to harbor two distinct human genomes - the version contained in normal cells, and an altered one contained in tumor cells.

But ours is not the only genome on the block. All organisms have genomes - not just humans and animals, but also bacteria, fungi, viruses, and other microorganisms that cause diseases. Studying microbial genomes as well as the genomes of their hosts (including humans) can shed light on the nature of infectious diseases. Moreover, analyzing the genomes of our closer relatives - primates, mammals, and vertebrates - and comparing them to our own genome can help researchers determine what parts of the human genome have remained unchanged over time and are therefore likely to be essential.

In 1990, researchers set out to sequence (determine the order of As, Cs, Ts, and Gs in) the human genome. The effort, known as the Human Genome Project, was an international collaboration that concluded in 2003. However, sequencing the human genome was just a first step - now scientists face the challenge of using the tools and knowledge gained from the Human Genome Project to better understand human health and improve disease diagnosis and treatment.

Want to learn more?

You can learn more about some of the efforts to decipher important information in the human genome and other genomes by reading about the Broad's Genome Biology Program. You can also read about how Broad researchers are applying genomics to the study of infectious diseases like malaria and tuberculosis by visiting the Broad's Infectious Disease Program page and the Genomic Sequencing Center for Infectious Diseases. In addition, scientists involved in the Broad's Cancer Genome Projects are working to document all of the genome-based abnormalities in tumor genomes.

The Human Genome Project website will give you more insights into the public effort to sequence the human genome. You can also watch a NOVA program on "Cracking the Code of Life" to find out more about the race to complete the sequence.

Interested in where the word "genome" comes from? Joshua Lederberg and Alexa T. McCray offer a brief history of the word, as well as a "lexicome" of terms ending in "-ome."

Read this article:
Genome | Broad Institute of MIT and Harvard

Posted in Genome | Comments Off on Genome | Broad Institute of MIT and Harvard