Feature: State of VFX in India


It has been 25 years since spielberg’s E.T. was released and sadly Indian cinema has not really been able to recreate that magic. It has nothing to do with technology, today Indian animators have worked on movies like Spiderman, Superman, Narinia, etc..According to me, there are two real aspects to it; one the vision of the maker and second is the ability. While there has been some improvement on the later, nothing much seems to have changed on the former. Take the case of much-hyped Roshan films (the first one was a copy of ET), they seem to completely take the ‘story-telling’ aspect for granted. It is all style, no real story. I really don’t know when things would change, we need someone like Peter Jackson, who completely revolutionised how VFX was used. He is a true magician put life into inanimate things through the use of computers. Quite sometime back, I had taken a stock of Indian VFX industry on CIOL, what is hapenning and what is possible. It is still valid today….
(http://www.ciol.com/content/news/2006/106012102.asp)
———————————————————————–

Will VFX arrive in India, Day After Tomorrow?


While Hollywood blockbusters like Day After Tomorrow saw a giant tidal wave submerge the Statue of Liberty, Indian films seem to be quite content to blow up a plane or two using computer effects…

Saturday, January 21, 2006


It’s evening, Ann Darow, a vaudeville artist and the giant ape are sitting on the tallest peak on Skull Island, gazing at the setting sun. The view from the top is serene and beautiful, a world that is almost magical, too good to be real.

But that world does exist, on one of the workstation at Weta studios in New Zealand, a ‘Matrixian’ sort of world made of 0s and 1s. Visual Special Effects — or VFX as it is popularly known — is constantly blurring the divide between real and surreal, nothing seems to be impossible anymore, as it was in the King Kong movie recently. Computers have brought the magic back in movies.

Right from the very start, filmmakers have been trying to make movies that defy conventional reality and in the process stretching the very limits of technology. George Méliès made the first sci-fi movie, A Trip to the Moon in 1902 inventing something known as trick photography.

The next big thing was the original King Kong made in 1932; Merian Cooper pioneered the use of stop-action model effects. George Lucas’ Star Wars in the 1970s opened the realm of possibilities with the use of robotics and computer effects. Steven Spielberg brought to life aliens in E.T., dinosaurs in Jurassic Park, sharks in Jaws and alien machines in War of the Worlds.

Finally, Peter Jackson went a notch higher, the Lord of the Rings trilogy proved what modern high-end computing can achieve. And if that was not enough, he put life in the giant ape King Kong. VFX in Hollywood is getting bigger and bigger by the day, every year big blockbuster movies are released that heavily rely on VFX to pull the audiences.

Bombay Dreams

In contrast, VFX in India is fairly primitive; we have barely achieved in Koi Mil Gaya, what Spielberg had in E.T. way back in 1982! Abhishek De, producer (VFX), Maya Entertainment, is of the opinion that, “we might be late, but we are catching up fast enough.” He cites the example of a film done by his company, Jajantaram Mamantaram, “It had over 90 minutes of special effects, fairly large by even international standards.”

Merzin Tavaria, creative director/VFX supervisor, Prime Focus, says, “VFX in India is coming up in a big way. Over the past 2-3 years, filmmakers are ready to experiment and explore new possibilities. These are ominous signs for the VFX industry in India.” Merzin talks about a project he is currently working on, titled Love Story 2050, he promises it will be a “real biggie” in terms of VFX used.

Color of Money

What are the shackles that bind Indian VFX artists from achieving the same as their brethren in Hollywood? “Cost,” says Abhishek, adding, “a visual effects sequence can cost anything from Rs. 5000 to Rs. 50,000 per second, or more, depending upon the complexity of the shot. Sadly, typical Hindi movie budgets are not in the position to commit that kind of money for VFX.” Thus, most Indian makers are content to just make their heroes do a summersault or leap from a rooftop with the help of VFX, or at best blow up obnoxious aircraft. All this work just about scratches the possibilities that are available with VFX.

According to Merzin it is has more to do with ignorance, “Directors in India are slowly waking up to the possibilities. We are coaxing makers to dream big and creating a market for VFX in India.” Recently, Prime Focus was in the news due to its association with a film Vaah! Life ho to Aisi that reportedly had a VFX budget of Rs. 60 million.

The Rise of the Machine

Are we lacking in technology or machines? Do we have the same hardware as ILM or Weta does? Pankaj Kedia, regional sales manager (South East Asia & India), Autodesk Media and Entertainment Division, assures that we do. “All the latest technology and software used internationally is available in India and the studios here are also adopting them. The Indian market is growing quite fast, in fact it is the largest growing worldwide market for Autodesk solutions,” he says.
Two-thirds of the top grossing films in Hollywood since 1993 have used Autodesk’s effects and editing technology; films like Day After Tomorrow, Lord of the Rings, King Kong, etc. Pankaj also believes that there is a growing realization of VFX possibilities in Indian makers, “Sanjay Gadhvi (director of Dhoom) is more convinced about the use of VFX after using it in Dhoom, so we will see a lot more effects in Dhoom 2. Similarly, Rakesh Roshan is also going all out for Krish after experimenting with VFX in Koi Mil Gaya,” he adds.
Yet, not every studio can afford Autodesk’s latest tech machines, an Inferno system costs approximately Rs. 30 million, a Flame system close to Rs. 18 million, a Flint system approximately Rs. 8-9 million. According to Pankaj, the VFX biggies in India are: Prime Focus, VCL, Prasad EFX, Maya Entertainment, etc.

A Man Apart

Weta Digital was just another studio in Wellington, New Zealand, a decade or so ago. But that was before a maverick maker by the name of Peter Jackson decided to rewrite history. He embarked upon one of the most ambitious projects of all time; bring to life J.R.R. Tolkien’s epic Lord of the Rings. Now, Weta is a VFX powerhouse, giving the best studios in Hollywood a run for their dollar. That’s the difference a single man’s will does. Do we need a desi Peter Jackson, who could dream big and then make it come true as well? “Of course, that would help,” says Abhishek from Maya, adding, “we indeed need makers for whom cost does not matter, only the vision does.”

The View from Beyond

Jesh Krishna Murthy has worked on films like Batman Returns, Lara Croft Tomb Raider, The Cell, etc. He is currently setting up shop in India, launching a company called Anibrain, targeting Hollywood and the Indian market. He is quite effusive on the subject, “VFX is not about machines or software; it has more to do with pushing the limits. I do not see many people in India doing that. And unless we really push ourselves hard, like developing software solutions, plug-ins, etc., we will never reach the level Hollywood has.”

Another person who is of a similar opinion is Jim Rygiel, three times Academy Award winner for Lord of the Rings (LOTR) trilogy. He was the VFX supervisor for the three LOTR films. He is an industry veteran, having worked in films like Ghost, Cliffhanger, 102 Dalmatians, etc. He is of the view that, Indian companies can replicate the success of Weta Digital, “however machines are nothing, so that makes all of the IT power in the world worthless unless you can get good artists to run those machines. The aesthetic needs to be addressed, much like the difference in aesthetics between an Indian film and a Hollywood blockbuster,” he says.

In conclusion, we have the machines, we have software, and we have the required talent. What we need now is the will: The will and self-belief to do anything.

Feature: CIA’s take on computer, IT, Internet and so much more…

Quite recently, CIA declassified a whole set of confidential documents, known as the ‘Family Jewels’. It is certainly not the first time CIA has done that, over the years there have been many such occassions. But this time, there was a lot of hype around it. Being an IT journalist, I was quite curious whether the gents at CIA were keeping a track of the way the humble computer evolved? So, I got cracking and was pleasantly surprised to discover that the agency really kept a close tab on the ‘wonder of computers’.
Not only that, I chanced upon lot of documents that detailed how the Soviets were failing in their IT designs, or how India was emerging as an IT powerhouse and why China would not really succeed in its modernization efforts. The problem was all these documents were scanned as images (at times quite unreadable), so I had to take a printout of quite many that seemed interesting (I have a fat dossier now) and manually type the text in.
It was quite an effort, as I was a wee bit worried, that some journo sitting across the globe at Time, or even BBC, might be working on the same concept. Anyways, after many ‘long nights’ the story was done and was published on CIOL. The story gives an idea of not only what CIA thought about computers , but also how it has evolved over the years. I am quite confident, that it makes quite an interesting read….(http://www.ciol.com/content/2170798471.aspx)
——————————————————————————

CIA on computers, technology and Indian IT

Culled from the declassified documents of CIA, here is a look at how technology and computers were shaping our world over the years. Of course there is lot of Bond, James Bond-like action involved…

Wednesday, July 21, 2007

The left-facing bald eagle perches atop a 16-point compass star. Metaphorically, the star denotes search for intelligence and information outside the dominion of the US Additionally, the compass star rests upon a shield, symbolic of defense and fortification. The emblem or the seal of Central Intelligence Agency (CIA), the premier surveillance and security agency of the US government, more or less says it all.

Shrouded in a cloak of secrecy, CIA carries covert work across different continents and countries. Not many know, what all the agency exactly does and there is little that the agency doesn’t, according to the few that happen to know and have spoken out. Established just after the Second World War, in 1947, CIA’s mandate is to obtain and analyze information about foreign governments, corporations, and persons. The agency or the company, as it is often referred to as, is also mandated to effect propaganda and handle public relations for the US administration. And finally, it is also, consigliere and caporegime of the reigning US President; ever ready to do his bidding.

Over the years, there have been many accusations and charges against the CIA, ranging from recruitment of Nazi scientists and SS officers to hiding information about UFOs; experiments on unwitting American citizens to concerted efforts to assassinate Cuban President Fidel Castro. There have also been allegations that the Agency had alliances with opium lords in Burma, Thailand and Laos; an assassination program in Vietnam; complicity in the toppling of Salvador Allende in Chile; recruiting and training Economic hit men, the arming of opium traffickers and religious fanatics in Afghanistan including a certain Osama bin Laden; the training of murderous police in Guatemala and El Salvador; and involvement in drugs-and-arms shuttles between Latin America and the US, the list just keeps going on.

But then, most of these is just speculation, and there is little proof of them. Unless and until the agency itself admits it, albeit obliquely. Over the past few years, the Agency, has been declassifying records under the Freedom of Information Act. These declassified records range from elaborate reports to short summaries on a range of issues, right from Cold War to Desert Storm. All these documents can be accessed from CIA’s website (http://www.foia.cia.gov/), using simple search functionality. There were quite a few documents, dubbed as the ‘Family Jewels’, were declassified recently.

Now, it is a well known fact that the agency has also been has been an avid technology user, it was even supposed to own quite a few U2 planes and reconnaissance satellites. Its obsession with technology is two-fold: one for its own use and second what other nations might be using to get an edge over the US Right from the onset, the agency has been making confidential reports and assessing technology. The big push for the same came in fifties with the launch of Sputnik satellite by comrade Stalin in the erstwhile USSR.

So, how could the agency miss the irrepressible computer, the miracle machine that was also evolving more or less around the same time? It certainly did not, on searching the term ‘computer’ on CIA’s website, one gets around 517 declassified documents, from the oldest on German textile industry using computers as a tool published in August 1945 to the latest on WMD search in Iraq dating from September, 2004.
Studying the documents is like taking a trip down the history, a proverbial time machine. One gets to know what really people thought about the wonder of the machine, what were the possibilities according to the people way back then and how things have changed.

A competent ‘mechanical slave’
In 1960, Joseph Becker authored a paper on “The Computer – Capabilities, Prospects and Implications”. The report talks about how “intelligence negotiates for the services of an obtrusive, demanding, but enormously competent mechanical slave.” That’s how the computer has been defined, a competent mechanical slave. First up, Becker distinguishes between two types of computers, one is the analog and the other is digital. The analog, he tells us, is more or less a mechanical device that “gives measurements of a continuum, notably time, direction, distance, or velocity, processes them mathematically as desired and displays the results in some measurable form.” A speedometer is an example of an analog computer.
A digital computer is like an abacus, he states. Digital computers deal with discrete numbers. “Numbers may be used to represent the layers of the alphabet or verbal symbols, the digital computer is the machine that has the major promise for handling the verbal data if intelligence,” he states.

The paper gives an instance of an digital computer, “Electronic computers like Remington Rand’s UNIVAC or IBM’s 700 series process numbers at speeds measured in millionths of a second, have an immense storage space or “memory” function precisely and accurately, and can process letters of the alphabet when these are numerically coded, treating them internally as if they were numbers,” states Becker as if he was wonder-stuck. In fact, it is this wonder that really strikes you. Don’t we get impressed when we see a humanoid like Aibo doing things, imagine what people must have thought way back then.

Becker talks about how there is a notion that “digital computers are endowed with near-human or even super-human qualities”. And discusses how a seasoned computer operator will argue that on occasion the machine has a ‘personality of its own’ and “his emotional involvement with the machine is such that research being done in man-machine relationships to arrive ay the right mix of human factors for happy and efficient work with a machine as colleague or subordinate.” It seems that the computer back then had a sort of halo to them. But he states emphatically, that the computer do not “think” but are driven through pre-determined set of operations.

In the future, Becker states, these computers would be extensively used in civilian and military work due to their high-reliability and faster processing. He lists out the different component of the EDP system as: input equipment, the computer, storage area, control mechanism and finally output devices. Becker also talks about a ‘recent innovation’ i.e., automatic programming. He talks about IBM’s Fortran language that makes a programmer’s job quite easier, as it contains only “38 statements”.

Becker also talks about how hardware technology continues to forge ahead, with the use of ‘cryogenic techniques and micro-miniaturization of circuitry’. “In the days to come very complex circuitry can thus be built into a cube of stacked wafers the size of a lump of sugar,” he says.

Yet, Becker’s prediction about the future of the digital computer is real attraction. “Looking ahead, designers foresee the day when refrigerated computers the size of portable TV set will operate on wall socket power. It is symptomatic that one of the serious design problems facing the computer engineers is that of minimizing the length of connecting wires, which becomes more and more critical as components get smaller and signal speeds approach the speed of light,” he states.

Looking back, one can say that he was not much off the mark, quite soon, the use of transistors curtailed the need of connecting wires and then of course there were the ICs. In the report, Becker further details the application of the computer and hints at how CIA could benefit immensely from computerization, namely through having a central digitized repository of documents. He also talks about the use of computers for automated translation of documents in foreign languages and to study photos in a better way. He also gives a sort of warning, “if we profit from the experience of industry during the past few years we should be prepared for some radical changes in the organizational structure as a result of the introduction of machines.” It could be that he was talking of labor substitution. All-in-all, the paper is quite detailed in its examination of the mechanized phenomenon, and he sure was not much of the mark.

The clandestine and conniving Soviets
One thing is for sure, after the Sputnik debacle. The CIA kept a hawk eye on all going on behind the iron curtain. Through out the years, there have been numerous reports and assessments on how well or how bad the Soviets were faring. Sometime, they would be quite premonitory in nature, talking about how USSR was rapidly deploying technology and specifically computers to good use. But more often then not, they debunked the Soviet efforts, it must have been quite a relief to the supreme commander at the White House.

One such report from 1977, titled “Soviet RYAD Computers: A Program in Trouble,” talks about how the Soviets were unable to succeed in creating clones of the IBM 360 series. “The USSR continues to experience serious delays in the development production, installation and effective use of its RYAD computers, which form the cutting edge of the Kremlin’s modernization program,” the author states. The reason according to the report was fairly simple, Soviet computers had been designed mainly for scientific applications and had insufficient internal memories.

In the report there is mention of how the Soviets had tired to clandestinely copying from the IBM 360 series. And also they were funding different hardware initiatives across Eastern European, namely in Bulgaria, Poland, East Germany, and others. Even then they were failing, quite miserably so as to say.

That was from the hardware point-of-view, another report goes on to state, how USSR was also trying to source Western software and computer applications that could be replicated at Moscow and other places. And just like the case in hardware, they were failing miserably in software as well. “The leadership of the Soviet software R&D community recognizes — as revealed in open-source literature – that the USSR’s software industry is technically backward. The Soviets know that, if this problem is not corrected, Western countries will be to extend their domination of advanced information-processing technologies for decades to come,” the author mentions.

Theorizing on possible reasons the author states that the lack of success with hardware might be the reason for the software industry. “The USSR is struggling with limited success to acquire and put into use the computer hardware that would be vital to a modern software industry,” it states. According to the report, to overcome this limitation, the Soviets were trying to source software from Western nations clandestinely. They were trying their hands at data sources codes, simulation software, algorithms for application development, utilities programs, etc. Later on, in another paper, there is a big hint that India might have a role to play in the technology transfer to the communist state.

India; an ally or a threat?
Americans were loath to the idea of the Russians (rather the Soviets) playing around with their machines or software. Under no circumstances did the US government want the Soviets to get hold of IBM machines or software of any kind. Thus for them, every State that had affable relationship with the Soviets was a threat. And India happened to be one such State.
In June 1985, there was a report published under the title, “India: A Growing Role in Technology Transfer.” Even after declassifying it, the agency has censored close to 70 per cent of the report (there is still a lot they don’t want the Indians to know). In the report, the author not only talks about the danger of India passing out the bits, bytes and ICs to the Soviets, but also in a small way captures the change that was taking place at that time.

There is a lot of talk of how prime minister Rajiv Gandhi was technolizing India and introducing a slew of reforms to that end. “Gandhi has introduced sweeping economic reforms that promise to encourage imports of sophisticated western technology and boost India’s ability to manufacture its own high-tech products. As a result, India will offer an increasingly attractive target for Soviet science and technology collectors in the years ahead,” the report warns.

It also talks about how after decades of slow and uneven growth, India had developed sizeable heavy industrial sector and an impressive economic infrastructure, and was poised for an unprecedented expansion into high-tech areas such as electronics and computers.

“We believe the scope for such growth is vast – India boasts a brand range of scientific activity and a large reservoir of highly educated and technically-trained manpower,” the author states. Adding that, “Gandhi is eager to make Indian high tech goods competitive in world markets, and the development of computer software exports is receiving particular emphasis.”
The alarm bells were ringing due to the close Indo-Soviet relationship. But there was an interesting observation as well, that even the Soviets might not like India to be more techno-savvy.“For the Soviets, India’s growing interest in expanded ties with the West and its quest for advanced technology present both a challenge and an opportunity,” the author mentions.
It seems that the American government had shared its concern with the Indian counterparts, as there is a mention about Indian security practices as well. “Indian officials believe their security procedures are adequate to deny the Soviets access to Western technology with the government, the military and public sector enterprises.

Six years after the report was published, USSR ceased to exist and the Indian economy was liberalized. Thus one could safely assume that the American fears might not have materialized or whether they actually did.

The coming of the Internet
The CIA was also tracking with interest the evolution of the Internet. The main concern of the agency was quite obvious: the Soviets could use it to steal information. A report “Soviet and East European Computer Networking: Prospects for Global Connectivity” published in 1990 list outs those concerns. The main concern for the agency was triggered by the entry in April 1990 of the USSR, Bulgaria, Czechoslovakia, Hungary, and Poland into the European Academic and Research Network (EARN). “The entry is likely to have a profound effect on scientific communities throughout the Soviet Union and Eastern Europe. The Soviet Akademset network, for example will become considerably more value if it provides a reliable bridge to vast foreign networks,” the author states.

There is also the story about the genesis of the Internet and talks about the “primary networks in the West”, and it sure is quite informative in a brief manner. “The best example of a government-subsidized research network is the Internet, which grew out of the US ARPANET project. Starting in the late 1960s, computers at the US Government, military, and commercial organizations conducting government-sponsored research were interconnected to permit the transfer of research data and maximize the utility of computers at various sites of the network. ARPANET spawned similar network, such as the US National Science Foundation-sponsored NSFNET to provide remote access to its supercomputer centers, and these networks themselves interconnected (inter-networked) – the result was termed the Internet. The Internet carries only unclassified information, and it now extends to many foreign countries,” the author states.

In the same report, there is an interesting story about what could be the first “primed” hacking attempt. Dubbed as the ‘Hanover Hacker’, the report details the Soviet attempts to hack US systems to gain information. The interesting thing was that they employed the help of programmers in West Germany (which was then an ally of the US) for the act.
“We have already learned about one Soviet effort to carry out illicit technology transfer via the Western research networks, the case of the ‘Hanover Hacker’. According to press reports, West Germans, acting at the direction of a Soviet handlers, engaged in a form of espionage over the Internet, Breaking into user accounts on computers in the United States and Western Europe by exploiting poor network security – including easily guessed or default passwords – the Germans stole proprietary or otherwise sensitive data. Although none of the data was classified, some of it could be considered ‘technical data’ that might have been restricted from export to the USSR by COCOM,” the author states.

But he seems to be quite baffled by the Soviet attempt. “According to the information provided to the press, the Soviets paid a considerable amount of money for the information and stolen passwords they received, The sum of the damage caused by the hackers – even bearing in mind the possible effects of the transfer of technology – does not seem to be substantial enough to warrant the expenditure. The operation’s exposure, however, could be considered a fluke (the hacking was not detected by any intelligence or police agency in any of the affected countries – it was discovered and pursued by a US academic researcher) and the Soviets may have anticipated far more productive future collection from the use of stolen passwords,” he states.
Beyond that, there is not much on the Internet, maybe there is oodles of reports that are waiting to be declassified.

Unmade in China
It was not only the Soviets that the Americans were worried about, they were also keeping an eye on the fledgling dragon, so as to say. China was also looked upon with distrust and was not thought to be capable of succeeding in its modernization efforts.

A report was published in 1986, titled, “China: Science and Technology Modernization.” The evaluation is quite pithy, that China though aspiring to be a big technology powerhouse could not really become one. “China has had little success in civilian applications of high-priority, advanced-technology sectors such as semiconductors, computers, and telecommunications. This failure will in our view hold back advancement in other high-priority areas, such as the development of automated production capabilities and communications networks for business and governments,” says the author.

Even then, the authors do note that energy and focus of the Chinese authorities in its modernization efforts. There is often the talk of how the senior government officials were quite serious about the S&T modernization and had set stiff goals that had to be achieved.

The goals were to commercialize technology, increase acquisition of foreign S&T, and finally breaking down of the barriers between civilian and military research and production.
But it was a tough task, according to the authors, as work conditions tend to be poor and key skills – such as computer specialists – were still in short supply. The big plus for the modernization efforts were the hundreds of Chinese students that were returning after being educated in Western universities. They were exposed to Western technology and could be a great help in the change.

Yet there was some amount of wariness. “China’s S&T modernization program creates both opportunities and problems for the United States,” The good thing was that there were already signs of a much more “open China.” But a powerful China could also be quite assertive. “China’s growing S&T capabilities and accompanying improvements will increase Beijing’s ability to project power against the nations on its periphery,” the author states.

There is the talk of how China may be able to narrow somewhat the gap in selected industrial technologies with the industrialized world over the next decade. “The degrees of narrowing is uncertain, but catching up fully with the West in any of these areas is very unlikely.” The general outlook for progress in high-priority areas was summarized as below:

Microelectronics: Gap will widen
Computers: Gap will widen
Telecommunications: Gap will remain or widen
Automated manufacturing: Gap likely to persist
Transportation: Gap will remain fairly uniform
Energy: Gap will narrow
Special structural materials: Gap will narrow
Biotechnology: Gap will narrow

“Of course, China will try and avoid becoming overly dependent for technological assistance on the United States or any other country,” the author adds.

In retrospect, one can safely say that the author (or authors) could not have been more off the mark. The transformation at China did go out pretty well, so much so that foreign firms are shifting almost all their manufacturing capabilities to China. The country no longer is dependent on technological assistance but is a global powerhouse of the same.

The world in 2015
Among the many reports, there is one that is quite interesting, titled “Global Trends 2015: A Dialogue About the Future with Non-government Experts.” The report was published in December 2000 and does crystal ball gazing into the future. It is quite an entertaining and speculative view of the future. As we are currently, somewhere in between (in terms of time), it still makes for an interesting read.

“The integration of information technology, biotechnology, materials sciences, and nanotechnoloyg will generate a dramatic increase in innovation. The effects will be profound on business and commerce, public health, and safety,” the report notes.
It also states that in the following years, the time between the discovery and the application of scientific advances will continue to shorten.

There are many observations on the political and security trends, economic dynamism, regional interaction, regional trends. But lets talk specifically about the IT front. “Local-to-global Internet access holds the prospect of universal wireless connectivity via hand-held devices and large numbers of low-cost, low-altitude satellites. Satellites systems and services will develop in ways that increase performance and reduce costs,” the author states.

The report goes on to list the countries that will play an important role in the IT revolution. “Among the developing countries, India will remain in the forefront in developing information technology, led by the growing class of high-tech workers and entrepreneurs. China will lead the developing world in utilizing information technology, with urban areas leading the countryside. Beijing’s capacity to control or shape the content of information, however, is likely to be sharply reduced,” it says.

“Discoveries in nanotechnology will lead to unprecedented understanding and control over the fundamental blocks of all physical things. Developments in this emerging field are likely to change the way almost everything is designed and made. Self-assembled nanomaterials, such as semiconductor ‘quantum dots,’ could by 2015 revolutionize chemical labeling and enable rapid processing for drug discovery, blood content analysis, genetic analysis, and other biological applications,” the author adds. There is also the mention of how by 2015, information technology will make major inroads in rural as well as urban areas around the globe.

But there is also a rider. “The rising tide of the global economy will create many economic winners, but it will not lift all boats. The information revolution will the persistence of poverty more visible, and regional differences will remain large.”
There is quite some thought that is given to Indian IT. Obviously, it must have been the success of the Y2K that must have woken up the Agency to the potential of Indian IT.

“India’s economy, long repressed by the heavy-hand of regulation, is likely to achieve sustained growth to the degree reforms are implemented. High-technology companies will be the most dynamic agents and will lead the thriving service sector in four key urban centers- Mumbai, New Delhi, Bangalore, and Chennai. Computer software services and customized applications will continue to expand as India strengthens economic ties to key international markets,” the author adds.

On China, the author is unwilling to bet, saying that, “estimates of China beyond five years is fraught with unknowables.” But he is emphatic on Russia’s decline. “By 2015, Russia will be challenged even more than today to adjust to the expectation for world leadership to the dramatically reduced resources it will have to play that role,” he states.

So this is how the world seems in 2015, India continues to go stronger thanks to IT, so does China, but not Russia. We are half way through; it remains to be seen whether in the next few years the authors would be proven right or wrong (they seem to have miscalculated Russian resilience, Vladimir Putin seems to be doing quite well for himself and his nation, as of now).

Computers @ CIA
After talking about computers across the globe, once chances upon a rare report that gives an idea of how the computers at CIA had evolved. In a report, “30 and Thriving” published in 1991, there is the talk about technology changes at the agency. At the onset, there were the AIWAC, which was a total-batch-processing system used for U2 measurements and worked on paper tape. It was the agency’s first digital computer. Next came IBM 407 and 1401, they generated “blip sheets” — massive paper printouts with up to seven carbon copies of historical data and blank data-entry forms. The 1401 were an improvement as they had four magnetic tape drives, and “8,000 bytes of memory”!

In 1962, came the Univac 490 systems, with 32,000 words of memory and were used for measurement and information procession. Further advancement came with the Univac 494 that brought the wonders of remote-batch processing. The agency went in for the Univac 1100s series in 1975 and in 1986, they were complemented with Sun terminals, it was then that the “light pens were replaced by the mouse.”

As of 1991, the year the report was published, the agency had two Uniysys 1100/93s’ and one Unisys 1100/91. “Together,” the author says proudly, “these systems offers 160 million bytes of memory (approximately 152 MB) and 146 billion bytes (approximately 135 GB) of storage and process 48,000 transactions a day.”

Another document honors Albert D “Bud” Wheelon, responsible for the creation of the Directorate of Science & Technology at CIA. Wheelon was mandated to create the department in a jiffy, post the Cuban missile crisis in 1962. In the report there is a quote by a person who eulogizes Wheelon and talks of how the research at CIA and other agencies have helped the progress of science.

“S&T research contributing to the processing and analysis of vast amounts of information which the agency collects daily. Research in improved data extraction, automatic database generation, machine translation and information retrieval – in multiple languages – have been major efforts supported by interagency groups, often led by S&T officers, and in several cases have led to commercial products. How else will the analyst or policymaker get the right amount of information in the form he wants on his desk when he needs it?” he says.

“The forerunner of the Pentium chip came into being because of some visionary officers in the S&T who believed in a radical concept – the RISC processor – worked with industry in the early 1980s to see that it got a fair chance – now look where we are,” he adds while giving other examples like tools for urban planners, 3D software, and facial recognition tools, etc. that have come from the agency stables.

So this is how the world of technology has evolved over the last years. Surely there are hundreds of thousands of documents and reports that have yet to see the light of the day. At least we can be sure of one thing after reading all these “Family Jewels.” There is so much more that we really don’t know. But we do know that the bald eagle is keeping an eye on all of us, making discrete notes.

Feature: On war and technology

Was reading about the BrahMos missile today, it is indeed India’s finest achievement in terms of warfare technology (the disappointments have been no less, ALH, LCA, etc.). Future wars will not be fought by men but by machines is a reality. Take Desert Storm as an example or even Kosovo or Afghanistan. Technology and war go hand and hand. All this reminds me of a story that I had done long time back, in fact when the troops had invaded Iraq, that touched upon the man and machine connect. The story was published in the Financial Express. Read on…..(http://www.financialexpress.com/fe_full_story.php?content_id=31145)

—————————————————
Tech Rules But You Still Need A Man To Fight The Battle
Some 500 years before Christ, Babylon near modern-day Baghdad was a bustling centre. It is often referred to as the cradle of human civilisation. King Nebuchadnezzar, after conquering Jerusalem, built a city of unprecedented charm complete with sturdy fortresses, moat, drawbridges, temple of Baal, tower of Babel and of course the Hanging Gardens, one of the seven ancient wonders. It was sheer technology, way ahead of its times, that created these monuments.

And ironically, technology is again playing a crucial role in Iraq, this time in its destruction. War and technology go arm-in-arm. The invention of gunpowder was responsible for the colonial ambitions of Britain. The U2 boats, torpedoes and the flying bombers stoked the Third Reich’s expansionist ambitions. And technology is once more in the limelight with Operation Iraqi Freedom, the satellite-guided Cruise and Tomahawk missiles, the Joint Direct Munitions (JDM)s, and the works.

From the 1991 Gulf War to Gulf War-II, there has been a distinctive shift in the way the wars are waged, a shift in favour of technology. Not only have weapons evolved in terms of destructive power, they are also smart now. For instance, Cruise missiles can hover around the target for an hour or so, waiting for a ’go’ from the command centre. The whole operation is being run from thousand of miles away in Virginia, US and Bahrain.

War is a logistics nightmare and technology comes to its aid. With the help of technology, there is much better co-ordination of the allied force, notwithstanding cases of ‘friendly fire’. It is now possible to manage and employ a variety of means of invasion — B-52 bombers flying down from the UK, Stealth bombers flying in from Diego Garcia, Apache helicopters from Kuwait, and missiles being fired from aircraft carriers in the Persian Gulf. All this is due to the intensive computerisation programme undertaken by the US defence department. There has been a seamless integration of sensors, communication devices and the weaponry systems in a single network. And this is the difference between the Gulf War in 1991 and the current one.

Says Lt General (Retd) Vinay Shankar, “In 1991 Gulf war, the weapons were in an experimental stage. Today, they are refined and calibrated. Take the example of the Patriot missile systems, they weren’t working then, but today they are bringing down the Iraqi missiles with amazing regularity.”

Any war is waged in three stages: surveillance, target acquisition and destruction. Spy satellites, UAVs (unmanned aerial vehicle), AWACS have been hovering over the Gulf region for ages now. The command centre has created a databank of targets like Saddam’s palaces, Ba’ath party’s offices, running into thousands, and every strike is collated with the data bank. Even the missile path for each Cruise and Tomahawk missile is charted on a computer keeping in mind the shortest possible route, traditional airline fly path and even high-rise buildings.
The other important aspect is target acquisition. In 1991, Saddam could fire Scud missiles at Tel Aviv, Israel, with impunity from mobile launchers as it would take hours for the allied forces to precisely pinpoint where it was launched from and then destroy it. But, now, with the AWACS hovering at 35,000 feet over Iraq, it is a matter of minutes before the launcher is located, and its coordinates sent to the nearest F-16 fighter jet on a sortie.
Says Air Marshal (Retd) VK Bhatia, “In the past, the pilot had to make an eyeball contact with the target before firing, thus factors like visibility and weather had a bearing on the strike. But with the current precision-guided range of weaponry, these factors have been made redundant. In fact, with the current GPS (global positioning system) guided missiles, the target can be changed at the very last minute.”

This has been made possible due to the new approach of the US defence sector – the way it is embracing proven and cheaper technology. Pre-1991, the establishment went in for exclusive and tailor-made systems. But now, they are adopting, cheaper and proven technology off the shelf.

“Operation Iraqi Freedom will change the way wars will be fought in the future. This kind of precision is unassailable, striking a target while buildings on either side are spared. In fact, I feel, it is the best advertisement for American armament manufacturers. No wonder they refer to the bombardment as ‘Shock ’n Awe’. These weapons are way ahead of their counterparts, even advanced nations like France, the UK and Russia cannot match the weapon systems of the Americans,” quips Jane Defence Weekly’s correspondent Rahul Bedi.

But technology comes at a cost. The precision munitions are as much as 30 times costlier than the ‘dumb’ munitions, but by virtue of their being accurate, they curtail the need for heavy bombardment.

Where does India compare with the military might of the American forces? “Nowhere,” says Lt General Shankar adding, “We are still generations behind, in terms of technology. The country will have to spend money in an intelligent manner to leapfrog into the big club, lest we be another Iraq.”

But, Air Marshal Bhatia says that the armed forces are embracing technology much more readily. Take the 1999 Kargil War for instance, where laser-guided pounding of Tiger Hills by the Indian Air Force helped in quicker expulsion of the enemy. The Mirages were able to pound targets due to the technological advancements made.”

The war has spilled over the World Wide Web too. During the previous Gulf War, Internet was at a nascent stage. Today, it is uniting people across geographies for and against the war. Blogs have become a major source of information. Take for instance, the blog by a supposed American soldier Smash (http://www.lt-smash.us/) reporting from the front. On the other hand, there is Salam Pax (http://www.dear—raed.blogspot.com/), a resident of Baghdad, who provides an insight into these turbulent times. Scores of US websites have been defaced by anti-war protesters, while pro-war hackers have ensured that Al Jazeera’s English site (http://www.english.aljazeera.net/) does not run at all.

But as the battle reaches the urban landscape of Baghdad, the technological edge enjoyed by the allied forces may get more or less nullified, feel analysts. The Republican Guards are more atuned to the weather conditions and the terrain, and then it will be a victory of the fittest.
A centimetre off-track on a computer becomes miles on the field. As the old military man says, you can batter the enemy with missiles, but you still need a man to fight the battle.

Feature: Microsoft & Novell pact

Proprietary Vs Open Source..is an interesting debate for any tech journo, so when Microsoft closed a pact with Novell last year, how could I resist penning my thoughts on this contentious issue….The story was published in the Dataquest Magazine & on CIOL.com and got quite many comments from people from all over the world….That’s best compliment for any journalist….(http://www.ciol.com/content/search/showarticle1.asp?artid=91191)

—————————————————

Divide to rule?

The Microsoft-Novell pact has really jolted the IT Industry across the globe. What are the reasons? What could be the implications? There are no real answers; only a few guesses and dollops of hope.
Raymond Noorda must certainly be turning in his grave. It has been barely a month since he left for the pearly abode and already his legacy has been undone. For over a decade, Noorda fought a relentless battle against the company at Redmond. Noorda was a former CEO of Novell and to him, William Henry Gates III was an unscrupulous usurper who needed to be stopped at all costs. Novell and Microsoft were bitter enemies, nothing less and could be more.
Thus, Microsoft came out with LAN-Man to beat Novell’s Netware and Novell went on a buying spree, for instance WordPerfect, to take on the might of Microsoft on the desktop space. Sadly, Novell wasn’t David and burnt itself hollow in its battle with the Goliath. In 1993, Noorda parted ways with Novell to establish the Canopy Group that invested in a whole lot of companies working in the open source space. Novell dragged on.
A decade or so later, Novell did a course correction and in 2003, jumped on the open source bandwagon with the acquisition of SUSE (a few months before acquiring SUSE, Novell had acquired an open source application developer company, Ximian). Despite the shift, Novell could never regain its past glory. It was a distant second to another open source major, Red Hat. That was the state a few days back till Novell decided to sellout.

Embrace, Extend, Exterminate
Since, the eighties, Microsoft has been at loggerheads with some or the all the IT companies. It is renowned for the subversive tactics that it employs to nullify opposition. “Embrace, Extend, Exterminate” is supposedly the corporate philosophy that it lives by. In its three decade of existence, innumerable companies have either been gobbled up or simply run out of existence. Gates (and now Steve Ballmer, the CEO) do not look kindly at competition.

Sun, Oracle, Apple, IBM, you name it, all have been detractors of Microsoft. Google was one of the few companies that was able to steal a march over Microsoft and establish itself as a leader in the online space. Yet, one of Microsoft’s favorite bugbears has been a product company with a cute penguin as its trademark, Linux. The open source movement is an anathema to Microsoft. The company propagates proprietary systems and is loathe to giving anything away for free or even opening itself.
‘Halloween documents’ is the name given to internal Microsoft memos that were leaked to the open source community in 1998. It is a revealing commentary on how Microsoft perceives competition, mainly Linux kernel-based operating systems. The memos dub open source software as ‘a growing long-term threat to Microsoft’s dominance of the software industry.’
The documents supposedly go on to acknowledge that certain parts of Linux are superior to the versions of Windows available at the time, and outlined a strategy of “de-commoditize[ing] protocols & applications”; or basing networks and documents around proprietary standards, thus they can only interoperate with machines that work on Microsoft OS. That was at the turn of the millennium.
From competition to coopetition
Noorda in his heydays had popularized the term coopetition, i.e., cooperative competition. This philosophy is the supposed basis on which the Microsoft-Novell pact is based upon. The pact has been touted as a symbiotic breakthrough. Yet, on closer analysis, there seems to be fairly little that Microsoft seems to be getting out of the deal. But then, remember what your kindergarten teacher repeatedly asked you to rote, ‘appearances can be deceptive.’

Novell is in a rag-tag shape; SUSE-Linux was certainly not a match-winner. The deal with Microsoft seems to be god-sent for Novell. First is the cash inflow, Microsoft would be paying Novell a sum total of $380 million, that includes payment for SUSE Linux Enterprise Server subscription certificates, as money for patent cooperation. Microsoft has also dangled the olive branch, it will not sue Novell’s customers for patent infringement. It will also market Novell’s Linux version to its existing customers.

The new friends would also collaborate in the development of modern technologies in the space of virtualization, management and document format compatibility (remember the Halloween documents). So now, Linux, importantly SUSE-Linux and Windows will be interoperable. A great victory of sorts for the open source movement, or is it?

Devil is in details
How does Microsoft gain from the whole deal? Is a question that is rankling many minds. Going by Microsoft’s track record, it should not be too hard to extrapolate. The open source movement was turning out to be quite a formidable challenge for Microsoft (the likes of Google, Amazon and other Wall Street firms were using open source systems), there was still some spadework to be done.

As of now, there are two main players servicing the open source market, namely Red Hat and SUSE-Linux (Novell). While Red Hat has close to 80 per cent share of the market, Novell makes up for the rest. There are a few other smaller distributors like Ubuntu, Xandros, Linspire and others.

Strangely, just a few days before the Microsoft-Novell announcement, Oracle had decided to market its own version of open source system, quite similar to Red Hat’s. Thus, Red Hat could give Microsoft a formidable challenge in the days to come. Now in one stroke, the open source market is divided in two camps: one blessed by Microsoft and on the other end are the baiters. While it is quite fashionable for open source developers to chant ‘Win-Down’ slogan, corporates and organizations would rather prefer a more peaceful and cooperative model. The preference for interoperable systems could boost Novell’s sagging fortune and eat into Red Hat’s share (embrace and extend).

“Once the details of the agreement are clear, I anticipate Red Hat will react in some way or the other. There could be more surprises in store, in the days to come,” Bhavish Sood, principal analyst at Gartner tells CyberMedia News. He also adds that there is not much clearance on the technology roadmap. “The goals are all good. What we await is a clear cut strategy roadmap, of how Microsoft and Novell will go about achieving it.”

The irony was not lost when Ballmer made a statement at the press conference, “we’re here to announce a set of agreements that will really help bridge the divide between Linux and Windows.” Did Linux really need a bridge that was built in Redmond? Meanwhile, Ron Hovsepian, CEO, Novell, talked about how he initiated the talks with Microsoft and how in the end, “this announcement gives our customers interoperability and peace of mind all in one.”
Indian speaking
According to analysts and market sources, Indian players are quite excited at the prospects of the future. There is significant support for open source systems in India and now companies could go in for heterogeneous systems, combining both Windows and Linux. “A majority of servers in India are already on the Windows platform, this would give certain users the liberty to go in for multiple environments, using SUSE-Linux,” says Doug Hauger, chief operating officer, Microsoft (India).

Hauger also pooh-poohs the ‘embrace, extend, exterminate’ talk. “All this talk does not really make logical sense. No one owns or controls GPL (General Public License), so where is the talk of exterminating it? It all seems quite humorous,” says Hauger.

He agrees that Microsoft could look at a broader initiative in the future, involving more players like Red Hat. “This pact has really broken new ground. What I find most exciting is that how mindsets will change in the days to come. The religious fervor sort of days (oh, I do not like Microsoft!) are over and been replaced with technical and technological talk. This is the evolution towards a mature marketplace, a place where, technology will take precedence over everything else,” he adds. Meanwhile, the Novell India team seems to be in a celebratory mood already. According to sources, the top management is currently in Paris for ‘official work’.
Sleeping with the enemy
The late Noorda had supposedly thwarted two acquisition attempts by Microsoft, after a failed merger attempt. If Gary Rivlin’s “The Plot to Get Bill Gates,” is to be believed, Noorda liked to refer to Gates as “Pearly” and Ballmer as “The Embalmer.” According to Noorda, Pearly promised the heavens, meanwhile Emballmer dug your grave.

Hopefully it is a different Microsoft and a different strategy. Probably, history would not really repeat itself. Just one final piece of advice for Hovsepian, when you dine with the devil, make sure you do not end up on the menu. May Noorda’s soul rest in peace. Amen!

Interview: Alex Burn (COO, WilliamsF1 Team)

Alex had come down to India sometime back to sign a contract with Tata Technologies, it was then that I had met him. He seemed quite eager and enthusiastic about his trip to India and I had suggested some ‘typical Maharashtrian’ delicacies, like missal pav. He couldn’t try it due to paucity of time, but promised that he will next time round. Am waiting to hear from him. This interaction was published in the Dataquest Magazine: (http://dqindia.ciol.com/content/cio_handbook07/GlobalCIO/2007/107022801.asp)
*************************

‘IT is making us a whole lot quicker… to work at speeds similar to our race models’

The date May 1, 1994, has a special significance in the sport of F1 racing. It was the day when Brazilian Formula 1 driver, Ayrton Senna de Silva died in a car crash in San Marino Grand Prix in Imola, Italy. He was racing for the Williams-Renault team and was in a winning position, when his car crashed into unprotected concrete barrier. The death of Senna brought the dangers of the sport into the limelight.

Post, 1994, all Formula 1 racing teams have put in massive security procedures in place to ensure that such a event does not occur again. IT plays a very critical role in this aspect, as companies are using the latest computational technology for better car design.

WilliamsF1 has been in the racing circuit for around three decades and is renowned for its FW models released year after year. The company recently signed an agreement with Lenovo, who would be one of the main sponsors of the team. Sometime back, Alex Burns, chief operating officer, WilliamsF1, had come down to India to visit Tata Technologies facilities in Pune. In an interaction with Shashwat Chaturvedi from Dataquest, Burns talks about how his team is using IT and why outsourcing might not be such a bad term after all. Excerpts.

What role does IT play in the development of an F1 car?
Technology is a critical element behind the success of any F1 team. Today, without any exception, every F1 team is investing heavily in latest tech mechanisms to get the best out of their models. Take the case of FW28, used in the 2006 season, we used over 4,500 CAD drawings during the design phase. We are heavily dependent on computational fluid dynamics, telemetry and other to not only develop an F1 car but also run it well.

To be frank, the FIA has introduced rules over a period of time that effectively slow down the car. It is driven by security needs, because otherwise all the cars would be trying to increase speed aggressively. There are whole lists of tests and reports that one has to complete before launching a model. There are the wind tunnel tests and crash analysis tests. Today, with the latest cutting edge applications, not only is it more cheaper than the traditional way but also a whole lot quicker. All this is only due to IT at work.

Once the model is up and running, what is the role that IT plays?
Running an F1 car is a highly data intensive job. For instance, over a weekend of grand prix race, close to 7GB of data is generated. This data needs to be meticulously analyzed and the design changes need to be implemented quickly. Also, this requires speed. At times, we work at speeds similar to our race models. When the car is running there are thousands of sensors that are attached all across the body reporting on different parameters. To make sense out of all this data and implement changes quickly, is a job that is best done with the help of IT.

We function at the very edges of technology. Our work is quite akin to the space industry. The components have a short life, we are constantly testing and incorporating changes. It is a very dynamic industry.

There is a general feeling that F1 racing has become overtly technology driven, the cars are more like computers. Your take?
I do agree that there is a general feeling of overuse of technology but you need to understand the reasons behind it. Since racing is a very dynamic and speed driven sport, any small error can be huge, not only in financial terms but also in terms of risk to the driver. Thus, one has to ensure all the safety and security that one possibly can, this is where IT is extensively used. And to that end, I support the use of technology. But, at the end of the day, the car is just an entity in the hands of the driver and it depends on the individual skills of the driver to steer the car to the premier spot.

What is the reason behind your engagement with Tata Technologies and the benefits of outsourcing?
The very same that are driving a host of companies around the world, namely time and money. Developing an F1 model is big money, and, as I said earlier, a lot of this is in the technology costs. Using the skills and facilities of companies like Tata Technologies we intend to shorten the development time and also decrease costs. We have a production cycle from September-March. That’s when we develop models for the next racing season. We would be working with Tata Technologies (Incat) on CAD models, etc for the FW29 model. Hopefully, as time goes by, we will increase our engagements with Tata Technologies. The quality of skill at Tata Technologies is high, and they have the ability and the wherewithal to put in the requisite numbers if need be for a project. It has the makings of a great marriage. We can do things faster and also in a cost-effective manner due to our association with Incat.

A word or two on the upcoming FW29?
In many ways FW 28 did not really meet the expectations of the WilliamsF1 team. We have learnt a lot from our outings in the 2006 season and we are going to apply them during the design for FW29. As we are retuning back to using the Toyota engine again, hopefully things will be very different in the coming season.

Interview: Noble Coker (CIO, Disneyland HK)

One would be hardpressed to find a person who does not know about Disneyland. But a lot many do not know that behind all that magic there is a lot of IT & Technicality. It was amazing interacting with Noble COker, who is the CIO of Disneyland HK, he is a man without airs and seems to love the challenges that keep cropping up all time. This interaction was published on CIOL, the link is (http://www.ciol.com/content/developer/newsmakers/2007/107031601.asp)
*************************

‘Magic @ Disneyland

Noble Coker marched into the meeting room. As the CIO of the upcoming Hong Kong Disneyland, this was his first interaction with his local team members. For the meeting, he had prepared an agenda and was ready to thrash it out with his colleagues. Strangely, throughout the meeting the team kept mum. On being pushed by Coker, the team members would only show their appreciation for the project.

“They found it incredulous to tell their boss that he was wrong, while it was much the case in the US,” he recalls. This was Coker’s first cultural shock, but he quickly learnt the ropes and next time round he was prepared with a solution.

To say that Coker is a fast learner, would in no means be an exaggeration, after all during his college days he learnt Lao language from all the refugees who were being resettled in California and later on took classes on the language, “to pay my way through college.” He joined PriceWaterHouseCoopers as an analyst and subsequently was hired by Disney as a programmer. Rising up the ranks, he took up the challenge to oversee the construction of the fifth Disneyland in Hong Kong, and the rest as they say is history.

In a freewheeling interaction with Shashwat Chaturvedi from CyberMedia News, Coker talks about the “magical experience” at the Disneyland and how IT makes it happens. Excerpts:

Can you tell us about the use of IT at Hong Kong Disneyland Resort? How much has been drawn from the other Disney parks in the U.S. and Europe and how has the HK Disney been unique in terms of technology adoption?
At Disneyland, the use of technology can be classified under four different categories, as follows. The first one is business transaction, the use of IT in hotel reservation, merchandise dale, food point of sale, etc. The second category is communication, the use of email, IP telephony for internal communication. The third category and a rather important one for us is the safety and security for our guests and visitors right from food to park monitoring. The fourth one is the entertainment; we use a technology in a variety of ways to enrich the guest experience. Hence, we extensively use IT at our parks.

Giving a precise figure on how the common technology between the Hong Kong and the international parks is a tough call, but I can hazard a guess that it must be around 80:20 international and local mix, respectively. We have used technology in very many unique ways in Hong Kong.

You have often emphasized on creating “magical experience” for the guests, can you share with some instances on how technology is being used to create that “magical experience”?
When our guests walk out of our park, I would rather have them remembering Disneyland for a great magical experience rather than a great technological one. I want people to say, “wow, how did they do this” and our technology is geared towards creating that very experience. Starting from the website, we have created a similar experience much like the one that a guest will experience at the park. At the park itself, we are using a variety of tech applications for instance, at the park there is a wireless broadcast that synchronizes the timing through the park, this ensures all the different elements work in perfect coordination, like the parade and the floats.

We also have piped music that is running through the underlying infrastructure. Or take the case of the newly introduced FastPass at Disneyland. In the past guests had to stand in long queues to be able to enjoy their rides. Now with Fastpass, they can register themselves for a slot later in the day and comeback in that slot and enjoy their ride. This helps the guests spend more time on the rides rather than queuing up. Even our park attractions use technology to reinforce the ‘magical’ feeling. Take the case of Stitch Encounter, based on the Disney character Stitch. He dynamically interacts with the guest and his responses are based on what the guest tells him.

One of the challenges (mentioned by you) was working with multi-cultural team, how difficult or easy is it to work with diverse teams?
To be honest, working with multi-cultural teams can be extremely difficult, if one is not sufficiently prepared for it. I committed a lot of mistakes and learnt through them. Such experiences forces one to remove our filters; filters that one acquires over time. The experience can be quite humbling. For instance, when I had come here, I was trying to achieve things without understanding the significance of different cultures. Typically, Americans have a bad habit of talking first, and listening later and giving away a lot content without much context. I quickly understood these issues and got down to working them out by understanding the people and learning more about their culture.

Do you think current-day CIOs pay much (more than required) attention to this aspect?
Though companies are going global, I still fell that as CIOs, we do not pay much attention to cultural sensitivity issues. By nature CIOs are naturally project driven and focused on getting the job done, and no one frets over such things. But, I personally feel, that resolving these issues can be critical to the success of a team.

What is the IT strategy roadmap for the future, i.e., technologies that are being tested for the future?
Going ahead, we have created a New Technology Group that has representatives from all the major Disney Parks like Hong Kong, France and the U.S. The group examines all the emerging technologies across the world and then uses them at the parks. For instance, some years back, the mobile penetration in Hong Kong was much ahead of what it is in the U.S., so we perfected the mobile applications out here and now they can be cross deployed in the U.S. market.

In the times to come, convergence across varying media will be big thing in the days to come. Today media is ubiquitous; there is a plethora of devices like iPods, mobile phones, etc. The challenge will be to deliver multi-dimensional experience. Guests in the future would want a more enhanced experience, so when they visit the Tarzan tree house, they would like more information on Tarzan or even like to see a movie clip of the film. We are gearing to deliver that enriched user experience. Disneyland will always be magical.

Interview: Scott Griffin (CIO, Boeing)

Scott Griffin is well-respected in the industry for the way he has turned the tables at Boeing. For the past few decades, the battle between Boeing and Airbus has been more than just 787s and A380s; it is war that will continue for as many years. And in such a scenario, the IT infrastructure can give a strategic advantage like none other. At least for now, Boeing seems to have it processes under control more efficiently than the European giant (which was beset by quite many technological issues in the past). Griffin is ensuring that Boeing does not give away the lead. My interaction with him was published in the Dataquest Magazine, the link being:(http://dqindia.ciol.com/content/cio_handbook07/GlobalCIO/2007/107041201.asp)
*************************

‘My biggest challenge is to speed up IT absorption to meet changing business requirements’

In 2006, Boeing overtook Airbus to become the world’s largest civil aircraft company in terms of orders. In many ways, IT has been the driving force behind the company with over $60 bn in revenues.

Founded in 1916, Boeing was the first airline company to employ digital technology to designing an airliner, and under the aegis of the current CIO, Scott Griffin, it is committed to technology. Griffin has been with the company for over two decades, working his way up through various departments. Currently, he is entrusted with the responsibility of not only ensuring that the 155,000 employees spread across the world are connected and productive but also that the company stays a step ahead in the face of onslaught from across the Atlantic, ie the Airbus A-380. Scott Griffin, CIO, Boeing, shares his experience and vision with Shashwat Chaturvedi of CyberMedia News. Excerpts.

Boeing as a company has been evolving over the last many years, especially so in the last few years-from an aircraft manufacturing company to being an aerospace and defense technology firm. How is IT being used in this transition?
Information Technology is the lifeblood of a technology company. It is used to create digital (3D) design of parts, plan tools and processes. It enables “design anywhere, build anywhere”, design reuse, design partner collaboration, globalization of the supply chain, and provides tools for increasing productivity and growth.

777 was the first commercial airliner to be designed using CAD. How have technology systems evolved at the company and how has IT been strategic to it?
777 was the first commercial, digital airplane. For the first time, we did not test the fit of the roughly 4 mn parts by building a mockup. We designed the parts and assembled them in the computer, using the Dassault Systemes CATIA CAD/CAM software. Then we checked their ‘fit’, using computer simulation.

Today, Boeing and our design partners do concurrent, 3D solid design of parts, plans, tools and processes. This has enabled us to take significant cycle time out of the design/build process, and will allow us to create derivative models with minimal effort.

With over 155,000 employees based across the globe, how do you ensure connectivity within the organization and what kind of IT infrastructure is in place?
Our infrastructure is global and standard. Boeing has customers, suppliers and partners in over 100 countries. The only way to provide a reliable IT infrastructure is to provide a standard IT infrastructure.

How has your role of a CIO undergone a change at Boeing?
Boeing IT has become a proactive partner in helping the Boeing business units achieve their growth and productivity objectives. Boeing IT employees support the Boeing enterprise, but we also provide revenue-generating IT through our existing programs.

Which component or technology (enterprise) will be take the major share of your company’s budget pie in the future?
Collaboration systems and infrastructure will continue to be a key investment for us in the next few years. Boeing has also been closely working with various Indian IT firms. How has been the engagement so far and can you also touch on all the work that has been done through Indian soil?In 1997, I became the CIO of Boeing Commercial Airplanes and within my first year, I had built partnerships with 5 Indian IT firms. We are still working with all of them today, and have added several more. My Leadership Team and I visit those IT partners in person at least once a year.

We intend to differentiate ourselves from our competitors in the way we use IT products, not in building the best IT products ourselves. That being said, Boeing has a healthy IT business selling IT products to our customers, and we will continue to build IT solutions where we see external customer need.


What is the IT roadmap for the future and can you touch upon some of the innovations that have been brought out by your teams at Boeing?
My systems strategy in support of Boeing is to ‘buy and integrate.’ Our intention is to buy commercial, off-the-shelf applications and integrate them into our architecture rather than write the applications ourselves. We intend to differentiate ourselves from our competitors in the way we use IT products, not in building the best IT products ourselves. That being said, Boeing has a healthy IT business, selling IT products to our customers, and we will continue to build IT solutions where we see external customer need.

Your views on the subject of cross-cultural teams?
Diverse and cross-functional teams provide the most innovative and timely IT solutions. Boeing is a global company, and our employees, suppliers and partners are diverse in terms of culture, nationality, and geography.

You have been associated with Boeing for over 2 decades, how has the journey been, and what would you term as your high-points and the biggest challenges faced?
I have had fun in dealing directly with airline and government customers. We are a customer-centric company, and it is challenging and rewarding to work directly with our customers. My biggest challenge is shared by my CIO peers, it is speeding the absorption of enabling information technologies in order to meet changing business requirements.

Feature: Whither Indian Software Products?

When will we have a Google or a Microsoft from Indian shores? Is a question that bugs me quite often. We are the undisputed leader in software services, but there has been little attention paid to creating IP or to be specific, software products out of India. It is a shame that only a handfull of companies in India are developing products; considering the kind of experience we have in the IT domain. But things might change, as these few players are getting ambitious by each passing day and their successes might herald a whole new shift to product development.. In the article, published recently in the Dataquest Magazine, I spoke about a small revolution that is taking shape, and how this just might be it..(http://dqindia.ciol.com/content/top_stories/2007/107051001.asp)

—————————————————
Point of Inflection
How some Indian players have sniffed out the global product opportunity, and are gearing up for the $350 bn market.

Thursday, May 10, 2007
Faqir Chand Kohli in not someone who can be easily ruffled. The octogenarian, fondly referred to as the father of the Indian IT Industry, has a very benevolent and benign stance towards things in life-right from the way the government is trying to enforce reservation on meritorious institutions to the way his created entity (TCS) is performing, or even how Indian politicians are not progressively inclined and just a bunch of nincompoops. But one thing surely gets his goat, literally. Call India an IT super power and suddenly one can catch sight of a stirring in his pupils. Kohli perks up and speaks in a voice that hardly seems to come from the grand old man of Indian IT.
“India is not an IT super power, how can you even call it so, we account for a few percent points in the total global IT pie. Look around you in India, the benefits of computerization have yet to really percolate to the commonest of the common. It is a fallacy to call ourselves an IT power house,” he virtually thumps the table.
Peering into his graying pupils one can discern a tinge of sadness. After the outburst, Kohli seems to calm down, accepting the inevitability of things and returning back to his calm and serene self. “Services alone will not make us a super power. We need to make our own hardware, our own software, our own applications,” he says unequivocally. And that is the truth, the bitter pill.
The Story So Far
For the past decade or so, we have been toasting the success of Indian IT; the flattening of the world or the emergence of Bangalore tigers. Year after year, Dataquest keeps coming out with the Top 20 volumes talking about how Indian IT super heroes, namely, TCS, Wipro, Infosys and others are faring. The most celebrated IT body, Nasscom too compiles an annual report that talks about the growth of the export market and many such statistics. The robust annual growth of the IT industry, some 30-40% year-on-year seems to have had a lulling effect.
Everyone hopes that the good times will continue to be and the million dollar contracts will continue to flow. It will, before the law of probability catches up or some other low-cost populous destination comes up, or a shattering innovation replaces the countless number of individuals employed in India. It is not an IT Armageddon, but a course of life. Many analysts and industry watchers have warned of the same, time and again. So what needs to be done? The answer has always been there, as Kohli said earlier, Indian companies need to look at creating IP, creating hardware products and, more significantly, making use of our intellectual capital, creating world class software products.
Consider this. India’s largest IT company, TCS, which is into consulting, services and business-process outsourcing, started its operations in the year 1968. Meanwhile, Microsoft was setup by a bunch of college dropouts in 1975, purportedly to sell software for the highly popular Altair 8800. This year, TCS crossed $4 billion in revenues and employs some 89,000 people globally. While, the Giant at Redmond (Microsoft) reported revenues of around $44 bn in 2006 and some 70,000 employees worldwide. This is how India’s largest IT company and the largest American (global, to be more precise) company compare.
And therein lies the answer. Indian IT companies have been primarily focused on application software development and implementation unlike global biggies like Microsoft, Oracle, SAP and their likes. Creating world-class products is the key to success. The good news is, India, and more importantly Indian companies, are discovering the benefits of pumping money into R&D of new products.
Changing Gears
Mumbai-based i-Flex solutions is a classical success story. It was in 1991 that Rajesh Hukku convinced Citigroup, where he used to work, to invest close to $400,000 in a software venture of a different kind. Rather than create application or software for foreign clients, he would make a product. After years at it his company, i-Flex, launched a solution for the banking industry, namely Flexcube. The product was a resounding success and found customers across the globe. It became so hot that Oracle decided to buy a 44% stake in 2005. A year and more later, Oracle has increased its share and is now the majority stockholder in the company with close to 83% stake.
There are more such stories emerging out of Indian shores, like 3i Infotech, Subex Azure, Cranes Software, Polaris, Ramco and others. “While last year was a great year from product companies’ perspective (product revenues touched $481 mn), and the winners were really the big firms such as 3i infotech, IBS, Ramco, etc. But the heartening part is that unlike the biggies, higher market growth came from players such as Tejas Networks, Ittiam, Tekriti Systems, Newgen, Nucleus, Skelta, and Axcend Automation, Aftek, and other companies that are often labeled as small players,” says TR Madan Mohan, director (Consulting, ICT Practice), Frost & Sullivan.
Little wonder then that more and more entrepreneurs are ready to take a plunge into this evolving industry. Take for instance the numbers given by Deepak Ghaisas, CEO (India Operations) and CFO, i-Flex. He is also the chairman of the Nasscom Product Forum: “According to figures available with us, there are around 346 companies in India that are into product development. Of this around 228 companies have a product offering.” Last year, the numbers of product companies were pegged at 250; this translates into quite a substantial increase in numbers.
Ghaisas provides another interesting insight into the numbers. “Close to 60% of these product companies have been started by entrepreneurs, mainly Indians returning from abroad who want to start something of their own,” he says.
Services Hangover?
Products is greatly different from services, is a phrase that every player utters at least once during the conversation. Indian companies have been renowned to take a de-risked approach to investment, and services is well suited for it. The product is a high risk and oft times high stakes business, the margins. For failure are pretty low. One has to invest in building a product over a few years and then market it to all and sundry. The RoI cycle can be pretty long.
Subash Menon, founder chairman, managing director & CEO of Subex Azure sums up the situation succinctly. “The product industry is yet to evolve properly in India. With the focus on export of software services most companies have ignored this segment, and consequently, there are only a few players in this space. Yet none can deny that the opportunity is quite huge and Indian companies need to work at making the best of this emerging industry,” he says.
Meanwhile, Amar Chintopanth, executive director & CFO, 3i Infotech seems to be a bit generous towards the services companies. “Over the years, the services giants have created a favorable atmosphere towards India. They must be credited for building brand India. Thus, product companies from India are no more taken as mere rookies anymore and are regarded with a certain amount of respect,” he says.
The Pot of Gold
According to market estimates, the global software product market is pegged at $350 bn and the Indian market is estimated to touch $7 bn by 2010. Thus, it is a big opportunity for Indian players both in the export as well as the domestic market. Unlike the services industry, products players have been known to hone their products in India and other developing nations before taking them to more mature markets globally. This is what i-Flex did in the nineties, 3i Infotech also followed suit, and so did Subex Azure and a host of product companies.
India is also attracting a host of companies who are setting up their development facilities. Take the case of PTC, the company has its largest R&D centre based in Pune. Meanwhile, last year, Nvidia had acquired a small Pune firm, Pace Soft Solutions. At that time, Jen-Hsun Huang, the CEO had said, “We have invested close to $50 mn in India and plan to invest close to $250 mn in the coming years.”
Not just BFSI
There has been a bit of an issue with Indian product companies so far. They have been mostly focused in the BFSI space. The reasons are plenty-the immense success of products like Flexcube and Finnacle could have spawned a whole generation of me-too players. Also the fact that till sometime back the only Indian sector that was able to provide business was the banking sector. Thus, there are a host of companies in this space.
Yet, there have been a few ventures that truly stand out. One of them is Cranes Software that makes statistical analysis tools. It has a unique business model of ‘Acquire-Enhance-Expand’. Re-engineering them to add new features and functionalities, and expansion to the global market in itself involves a significant amount of R&D. Newgen is another noteworthy example in the document management space and many other ventures like these.
Big Services Daddies
It is not only the dedicated players that are eyeing this segment, so are the big daddies like TCS and Infosys. TCS, after its acquisition of FNS has been relatively active in the core-banking space. It has not really been worth too much in terms of revenues for the company. As TCS is getting much money from servicing clients, it does not seem to be too keen on the products game.
Infosys has been accelerating on the products domain. Its banking solution, Finacle, enjoys quite a good installed base in the industry and yet its contribution to the overall revenue is not something to sing about. “Finacle contributes approximately 4% of Infosys’ total revenues. However, this data point does not reflect the fact that Infosys has packaged software only in the banking solutions space while services focus continues to be across verticals,” says Sanat Rao, global head, Finacle Sales, Infosys Technologies.
The issue is of mindset, it is quite hard for a TCS or Wipro or Infosys to break the shackles and succeed in the new domain. As long as the services industry is going great guns, the big daddies will continue to remain marginal players.
Governmental Interference
Most of the industry players complain about how the government has done nothing for the industry at large. By imposing excise duty on packaged software in the last budget, the domestic market has been badly hit.
“It is the IT industry which created the culture of quality, globalization, technical education, building world class infrastructure and a brand to recon with in the international markets. I can bet that if the government had consumed all the tax benefits instead, India could be a laggard country as it is in all other sectors. Minus, the rise and shine of the Made in India brand of the software and the BPO industry globally, India has no international standing whatsoever,” says Hanuman Tripathi, MD, Infrasoft Technologies.
Menon from Subex wants organizations like Nasscom to play a more proactive role. “They really need to promote the industry like they have done for BPO. Once the word is spread, more and more entrepreneurs will come forward and the industry will flourish,” he says.
Kohli’s Gift
Coming back to the father of Indian IT. Kohli has not let age or anything come between him and his vision. He was responsible for recently engineering a CBFL (computer based functional learning) method that has provided literacy to many thousands in rural India. He even talks about creating a product for the numerous kirana stores dotting every town and city in India. “These stores with a computer and a custom made solution could then compete with the Walmarts and HomeDepots of the world,” he mentions.
Vision is the key. If an octogenarian can still think about creating products and solutions that can be used by millions, what really encumbers millions of computer geniuses and management whiz kids from doing so? It is perfect time to move into the product space, to do something truly path breaking. The product industry is indeed at a point of inflection.
***X***
Talking Services
For over a decade now, i-Flex has been famous across the globe for a single thing, a core-banking product known as Flexcube. The product, developed in the early nineties, has blazed a trail like none other. Even today, though the company has a variety of successful brands and products, Flexcube continues to be a dominant force. And one of the chief reasons, according to Deepak Ghaisas, CEO (India Operations) and CFO, i-Flex Solutions, has been its adaptability to different cultures and market dynamics.

“We have been able to do so through the service support that we provide to the customer. We do not sell a product but sell a solution,” he says. And all the while his cash registers keep ringing. Services today reportedly account for around 50% of the company’s revenues. Not small considering that i-Flex’s revenues is estimated to be in the range of $330 mn annually.

“Services is crucial to us, not only from the revenues perspective but a lot other ways as well. We have found that services can often act as an incubator for the product, nurturing it in the initial phase. It can also be a very good employee retention tool as the turn-around cycles in a product company can be quite long,” adds Ghaisas.

As part of the initial strategy the company had deliberately avoided going all out in the advanced economies, instead it went to developing economies in Asia and Africa. Now, i-Flex is taking another step to ensure its continuous success. The service profile of the company is improving with each passing day, as it ramps up for the next level of growth. With Oracle’s (as it holds around 83% share in the company) marketing and servicing might behind it, i-Flex can truly change the way Indian product companies have fared till date.
Vision is the key. If an octogenarian can still think about creating products and solutions that can be used by millions, what really encumbers millions of computer geniuses and management whiz kids from doing so? It is perfect time to move into the product space, to do something truly path breaking. The product industry is indeed at a point of inflection.
Going Glocal
Ambition is a good trait, but like any overdose, can be quite hazardous at times. There have been quite a few cases in the recent past wherein a company that was cash rich expanded rapidly and burnt itself out rapidly as well. As Alexander Pope had once said, the same ambition can destroy or save.

3i Infotech believes in taking steady but firm steps and part of its strategy is to go glocal. With around 80% of its revenue coming from overseas it is pretty obvious that 3i Infotech needs to have a presence in all the countries that it works in. But it can be a costly and risky affair. Opening an office in a foreign country means investing precious capital that could have been otherwise used. 3i Infotech has taken care of that problem by going global through the local way, ie, through appointed local partners.
“As we have around 300-400 customers and most of them based overseas, it naturally makes sense for us to expand overseas. Thus we have adopted a partner strategy. For instance, before venturing into any country, we conduct a thorough research on the market and its potential. Once convinced, we appoint a partner in that geography and operate through him or her. As the business expands we add on a few partners more, and after a critical stage, we ourselves enter the country,” says Amar Chintopanth, executive director & CFO, 3i Infotech.
Currently, 3i Infotech has a partner network in more than 10 countries of the total 50 that it operates in. By using a partner, the company saves a lot on the cost of capital that would otherwise need to be invested. “And this capital we plough back into R&D,” adds Chintopanth. For product companies in India that are short of cash, going glocal is the best possible option.
Inorganic Approach
Subash Menon is a man who should be truly admired for his gumption. An electrical engineering graduate from a university in Durgapur, Menon decided to float a company in 1992, without much help or experience. He had an idea and the urge to make it happen. Subex Systems evolved from being a telecom SI to a product company focused on the telecom space.

Then last year, Menon decided to go full steam ahead. In a move that surprised many, Subex acquired UK-based Azure Systems for close $140 mn in an all share deal. At that point Subex was worth some $25 mn compared to Azure at some $31 mn. The new entity, Subex Azure, was well suited for the telecom OSS market. But Menon is an ambitious man and recently went in for two more acquisitions in the range of $100 mn. So how does the organic strategy works?
“M&As are an important part of our roadmap and we pursue both organic and inorganic routes to enhance our product portfolio. As a policy, we work on a 4-year roadmap, it clearly states where we want to be in 4 years time and how. The recent acquisitions are based on the plan that we have chalked out for 2010. In the last seven years we have made seven acquisitions amounting around $320 mn in cash and stock. We are also in the process of raising around $200 mn by issuing Global Depositary Receipts (GDRs),” says Menon.
According to Menon, Subex Azure will continue to look at expanding the inorganic way, and is looking for possible buy-outs in three areas, namely revenue maximization, service fulfillment and service assurance. “We have evolved being a fraud management solutions company to being a telecom OSS vendor. Our aims have become bigger and so has our addressable market. We intend to go full steam ahead,” says Menon.

Pola-rising Market
Sometime in 2001, Polaris did a reality check. It was established in 1995 and doing reasonably fair for a services company, but Arun Jain, CEO, Polaris, knew that it would not be able to compete with the likes of TCS and Infosys. For all its efforts, it would be tough to break into the big club. It was around this time that Polaris changed tracks. It adopted the Blue Ocean strategy; instead of slogging it out in the highly competitive services domain, why not coast along in the relatively newer space of product development. The company’s expertise in the banking domain would also come very handy. But even the banking domain had a few strong players like i-Flex and others. There were quite a few players competing on the plank of technology and cost. Polaris decided to bring its technical expertise on the table, and introduced componentized products based on SOA principles.


“The idea was fairly simple, but complex at the same time. Rather than selling a product, we decided to present a platform to our customers, whereby he or she could pick and choose modules or applications that were required by the business rather than going for a big-bang implementation. We termed it as Non-Disruptive Measured Steps Method or NDMS,” says Jaideep Billa, CTO, Polaris Software Lab.
With NDMS, companies were able to migrate from another core-banking platform to the Polaris platform with little or no hassle. And the results were there for all to see. “Today, top 7 banks from the top 25 use our solutions in some way or the other,” says Billa. Though Polaris could not be a shining star in the services domain, it certainly emerged as a force to reckon with in the product domain.

Does it Tally?
Till a few years back, Tally was the poster boy of Indian IT industry. A homegrown solution aimed at the small domestic players, Tally really grew in real stature. Since Tally package was customized for Indian needs and requirements, it had completely dominated the SMB space.

Over the years, the big ones like the SAPs and the Oracles of the world were focused on the big companies in India. That was till a few years and now the very same international have woken up to the immense opportunity in the SMB space. Suddenly, Tally was under attack with international players products at lower price. Its price plank was removed, the growth was stagnant and suddenly the company seemed vulnerable.
To counter the situation, Tally is trying to reinvent itself. After being funded by Reliance Mukesh Ambani Group, that also picked up a stake in the company, Tally has been trying to reach out to different markets like the Middle East. It has also decided to take the game to the enemies’ court, by venturing into the ERP space. Tally also came out with a solution for the retail sector. Time will tell if Tally will be able to tally all the different things that it seems to be doing or will it, just not tally.

——-EOM

Interview: Sam Coursen (CIO, Freescale)


Sam Coursen is a veteran in the IT industry, over 36 years — thats more than my age, thankfully :)– in an interview the FreeScale’s CIO spoke on whole lot of things that not only affect his company but also the overall industry at large. I liked his take on the onshore-offshore debate. His response, “..in the globalized world, everything is onshore somewhere”. Read the interaction as it was published in the Dataquest Magazine, the link being:(http://www.dqindia.com/content/cio_handbook07/GlobalCIO/2007/107062701.asp)
*************************
‘With IT, one needs to be agnostic in the outlook’

‘Am joining a 70-year old startup’ is what Sam Coursen had said while joining Freescale Semiconductors as vice president and chief information officer in 2005. And he was not much off the mark. Freescale had existed for quite many decades as part of the Motorola group; it was only in 2004 that the company branched out. Today, it is one of the largest semiconductor manufacturer, and a leading supplier to the automotive industry.

Even in these ‘three’ years of its existence, a lot has changed at Freescale, from being a part of Motorola to being bought out for over $17 bn by a consortium led by Blackstone Group LP. The Freescale purchase, which closed on December 1, 2006, is reportedly the largest private buyout of a technology company and one of the ten largest buyouts of all time. Sitting at the Austin HQ, Coursen is keeping a tab on everything, not only ensuring that everything goes according to plan but also keeping his eyes on the horizon for newer technologies and breakthroughs.

Thirty-six years is a long time, and that is how much Coursen has spent working in the tech industry, starting with AT&T and then Bell Laboratories. For seven years he was at NCR as the CIO, before he took up the challenge of joining Freescale. The best thing that this industry veteran likes about his job is that: ‘it is never the same. Technology is changing all the time and it keeps me on my toes…Moore’s Law has never stopped working and hopefully would not in the near future as well,’ he says with a flourish.

When Coursen is not in office or trotting across the globe, he can be found at a golf course, swinging his clubs. More recently, he has developed a liking for whiteboard surfing on a lake that is nearby. Coursen likes to travel a lot and loves Sydney, though he also likes Scotland, Germany, China and, of course, India, probably due to culinary reasons. In an interaction with Shashwat Chaturvedi, Coursen spoke about the changing role of the CIO over the years. Excerpts.

At the time of joining Freescale, you dubbed it as a 70-year-old startup. What would you term as your biggest challenge at Freescale?
When I came here, I found a very good team of professionals who were exceptionally good at their work. My biggest challenge was to get the team sort of pointed to overall strategy for IT, making it more holistic in nature. To that end, we have defined a vision state that we want to attain, and have gone for major projects. It was like adding a strategic level to the function. Fortunately, for me, the team was quite capable of actual execution, so it was not that big a challenge.

While joining a big MNC that is spread across different geographies, what according to you does a CIO need to bring to the table?
Several things, but first I think a CIO needs to have the experience of operating globally. One needs to be empathetic to differences in systems and cultures, and to succeed it is imperative to adopt a global sourcing attitude. In today’s world, the concept of onshore or offshore does not make a difference for a global company, because everything is onshore at least somewhere.

How have the systems evolved at Freescale, and what are the changes that you are bringing in?
Even when Freescale was a part of Motorola, it existed as a separate business unit, so the systems were pretty much independent, and at an enterprise level. That was quite good in many ways. Some of the systems or functions were integrated to the Motorola one, so Freescale had to actually implement them from scratch, for instance, the HR function was shared with Motorola, so in the first year of separation we went for a complete implementation of the SAP HR suite, one of the most complete implementations of the suite. So, we have one single instance of HR for all our employees spread across the globe, and all our HR systems are in one place and uniform. That is not common, I know from experience that a lot of companies have HR systems based on countries or location, which makes it difficult to do a complete analysis at times, for instance, salary planning or headcount, etc. Thus, we had a set of systems focused on the business independently even when we were at Motorola, and we had to augment the systems that were shared with Motorola.

Coming to the second point, we have initiated a number of projects. My personal favorite is enterprise data warehousing. It is a big thrust for us. While our ERP is SAP, we did have a few legacy systems like in manufacturing, while some systems were in Oracle. So, there existed these data islands that one could not do much about. We are in the process of creating architecture to bring all that data in one place. It is a long drawn process, but we are making good progress. Currently, we have the core data in place. We also have a number of projects, wherein we are replacing legacy systems with new ones.

Freescale has operations in around thirty countries and more, and an employee base of 24,000. How do all these people collaborate and work?
We use the same collaborative tools that are available to all companies like Net meeting, Live Meeting, audio and data conferencing, etc. The good thing is that we have a good culture of taking advantages of those in this company, so a lot of this collaboration happened over the IM or the email. Going forward, we are looking at modernizing product design, so we are implementing PLM suites, where in there are lot of cutting-edge tools for collaboration. We are taking the first step toward PLM by implementing product data repository. The current project is to put in the foundation of the data management piece, and then look at different collaboration tools that would ride on top of the database and allow us to collaborate inside the company as well as with our partners.


Obviously, being a semiconductor company, IP protection must be a big issue. What’s the latest on that front?
We are actively looking at enhanced technology and evaluating the potential. We do have a number of security capabilities in place that relate to data storage, especially sensitive data; we use tools like encryption and what not. The employees are also sensitized to not have data in environments where it is harder to enforce patents and copyright laws. In the future, there is lot of work happening on IP protection, so you have an emerging technology that allows you to protect by particular document. For instance, a company, Liquid Machines, lets you designate certain IP with every document, which will provide enhanced levels of security. We are looking at whether we can implement those, as these technologies are not inexpensive.

For seven years you were the CIO at NCR, before coming to Freescale. How has the role of CIOs evolved?
I must admit that it has undergone quite a change. I keep reading all these debates in Harvard Business Review whether IT is really a strategic asset that can enhance value of the company or a cost to be reduced. And, that debate relates to the job of the CIO. Let me explain, A CIO cannot afford to ignore either of those two views. IT has become pervasive in every company, and it almost underlies every process. That’s good news for CIOs, and bad news as well. The good news is that IT has become an important function that can make a difference; the bad news is that because it is so pervasive, it is in a way expensive as well. As it is such a big part of the expense of the company, CIOs need to continually look at how to do the IT function more and more efficiently. We need to focus on how the whole thing can be done in a cost effective way, thereby making the company more productive. For instance, invest in manufacturing capabilities, invest in data warehousing, thereby enabling managers to make faster decisions, better ERP, supply chain, etc thereby optimizing everyone’s process. So, I got to focus at least half of my time on driving IT to be very efficient, so we can have the most possible affordability to really drive business process reengineering, and get strategic advantage over our competitors. A CIO today needs a balanced approach to the job.

What would you term as your biggest nightmare?
I share the same fears as my other CIO peers, namely a complete network breakage, data center is going to burn up, etc. As most of the companies operate on top of the IT infrastructure, it has to be there 7X24X365. It is a challenge to keep everything running smoothly, as the world is getting more dangerous with the bad guys trying to break into your systems. If we are well connected with the rest of the business, I think we will build the right systems. But, to keep them running without a glitch is the challenge.

How is it working with a multi-cultural team, and what are your learnings?
There are a lot of learnings; to have people from all across the world with different perspectives can be quite advantageous if you can leverage the multiple perspectives. I visit different regions where we have offices to meet my teammates, and then have them visit me at the HQ, so that you get to know the people face to face. It’s good that companies today tend to drive standard processes globally. But, they do vary to some degree based on local culture, local laws, and local customer needs. One needs to have a good appreciation of what can be standard, and what needs to be truly effective, and the only way to do that is to work in a multicultural environment.

What are your views on outsourcing and the big debate on core vs non-core?
Core and non-core are things that apply at a company level. For a company like Freescale, R&D would be the core. But, with IT, one needs to be agnostic in the outlook. I strongly believe that if someone can do a particular function better than me, then I should let him do it. What I don’t really like is this big second company do a total outsourcing kind of a deal. Because, I really think that it is very hard for one entity to do all the things in an efficient manner. You could term my attitude on outsourcing as more like out-tasking. For instance, 20 years back we would have gone and bought our own PBXs and built our own voice networks. But now, you go to AT&T, or any other company and buy software, etc to set up a network. This works out quite well, so no one in the right mind would look at building a voice network by them. I see a trend toward specialization, but I want to have control on the different pieces that I outsource or out-task. The people who have regretted their actions are the ones who have taken their entire thing, and outsourced it to an EDS kind of a thing.

What is the strategic roadmap for IT at Freescale?
It is interesting, systems support not geographies, not organizations, but they do support processes or a function right? For eg, a set of systems support HR function, supply change, financial, etc. Based on this, we have organized systems by the processes that they support. I have created a structure where we have defined a function map of the company, there are like twenty of these functions like finance, HR, etc. We have identified a business owner and an IT owner of each of these twenty functions who work towards establishing the vision state that I had mentioned earlier. We are currently very focused on data warehousing and manufacturing excellence, we also have a lot of work on supply chain systems. There is lot of work to keep us busy in the future.

Now that Freescale has been bought out, do you see any changes in your function?
Not really, nothing much changes for me. The owners have changed, but we always had someone owning us. Also, I see the potential as quite positive. For a semiconductor company economies of scales are important, and if by being part of a big group it could be leveraged in a better way.

Feature: Story on an innovative company in India


This story was published in the Dataquest Magazine and is about an amazing couple from the US, the Binfords, who decided to not only make India there home, but also dstart up a venture in India that promises to revolutionize the way devices read human inputs…. (http://www.dqindia.com/content/Innovator/2007/107031701.asp)
*************************

Reading a Bright Future

Towards the late nineteenth century, William John Warner, a young Irishman, traveled extensively through the length and breadth of India to study the ancient Indian art of astrology. He purportedly visited a lot of Brahmins in North India, studying extensively from the ancient manuscripts possessed by them. Later on, he went to create his own sect of astrology, and was known to the world as Cheiro or Count Louis Hamon.

A century or so later, their was another person who landed on the Indian soil. But this time he was not here for knowledge, rather fashioning it from here. It was the time when the ‘world had not yet been flattened’ and ‘tiger was just a cub’, when Prof Thomas O Binford decided to shift to India bag and baggage. For over three decades, Binford, a leading researcher in image analysis and computer vision, was at the Stanford University as a professor of computer science (Research). In 2000, when he retired from Stanford, Binford decided to don the hat of an entrepreneur and work on a startup. Since, he would be funding the startup from his retirement fund, he had to be quite discerning about how and where he invests the money and that is where India comes into the picture.

Handwriting Recognition Engine
Around the same time India was emerging in the global consciousness, thanks to the Y2K success. Binford decided to take a risk, and in 2001 shifted to India with his wife (a former manager at HP). Settling down in Bangalore, Binford floated Read-Ink Technologies in the space of handwriting recognition. Read-Ink is currently building handwriting recognition technology that could be used in tablet PCs to mobile PDAs that use the stylus. A whole lot of companies across the world have been trying to solve this issue. “The case is fairly simple, every individual has a unique handwriting much like his fingerprint. How can computers decipher it? How can it be digitized? We are trying to solve this problem,” says Binford.

The company is close to building a high accuracy recognition engine for handwriting and printed text. Based on this, organizations across the globe can automate their data capture process. According to Binford, the engine is going to be launched in the market soon, and already has generated considerable interest in the market.

“We are currently in the prototype stage and working on the alpha version. Hopefully by the middle of the year the product will be launched,” he says. Read-Ink has close to 20 engineers that are currently working with Binford. “Most of them are IIT graduates, highly motivated and raring for a challenge,” he adds.

For the professor the biggest source of support has been his wife, Ione. She is the CEO of Read-Ink and takes care of all the operational issues, like managing the team, marketing, etc. She even runs a small all-night accounting back-office service for American customers. “It is a small service with few clients. It helps us to cover the cost,” she says. Surprisingly, Ione does not think much of the cultural shift. “Since, I am basically from Brazil, India seems quite like home, except probably the infrastructure is much better in Brazil,” she adds gingerly.

A few years into the venture, the Binfords made acquaintance with Pankaj Jain, a Stanford University and IIT Delhi alumnus. Jain was just back from a stint in the fabled Bay Area and was planning to do something on his own. He had earlier worked with Oracle and Seibel in the US. It was after the interaction that Jain and the Binfords launched another venture, Taxila Labs.

Named after the ancient university in modern day Afghanistan, Taxila Labs is working at developing a solution that makes vast amount of information available on mobile phones both through voice and data, thereby making mobile phones as potent as the PC.

“By using the technology developed by us, companies like retailers and service providers can build storefronts and portals that consumers can access on mobile phones and discover the relevant items with ease. These storefronts and portals provide the same choice that is available over the web through a PC and thus enable the Long Tail economies on the mobile phone,” says Jain. The company is very close to finalizing deals with a few companies in India and would be commercially launching the product soon.

“We intend to run and test this technology first in the Indian marketplace, before taking it to advanced markets like Japan and Korea. Considering the mobile explosion in India, if the technology works well in India, it can work anywhere in the world,” adds Jain.

Till date, the Binfords and Jain have been funding the ventures from their own pocket. They live and work in the same building, saving on rent. The ground floor contains a kitchen and employee dining room as well as the Binfords’ bedroom and employees’ guest rooms. Things might have been a wee bit tough, but they are sure that the technology that they are developing is going to take the market by storm, the way Google and Netscape had done in the Silicon Valley. Like Ione says, “I see a dim light at the end of the tunnel”. The distance seems to be decreasing and the light is getting brighter. Who knows, as in the twentieth century the world learnt palmistry thanks to India (merci Cheiro), probably in the twenty-first century the devices might learn to read handwriting, again courtesy India (merci Binfords).

Interview: Jonathan Palmer (CIO, IMF)


In an interview IMF’s CIO Jonathan Palmer spoke about how IT has made the organization more proactive and transparent. This story was published in the Dataquest Magazine, the link being:(http://www.dqindia.com/content/cio_handbook07/GlobalCIO/2007/107062201.asp)
*************************
‘IT has made us more transparent and accessible’

It is hard to come across a person who has not heard of the International Monetary Fund. Even harder to come across one who does not have a view on IMF (in all probability a negative view tinged with apprehensiveness and distrust). But the irony is that not many individuals seem to really know what really the IMF does. Often, one comes across new stories that belabor the IMF’s Keynesian approach and its frequent ‘austerity programs’. The Fund, as it is referred to as, is often chastised for it recommendations to poorer nations on currency devaluation, and balancing of budgetary deficit.

That said, the IMF plays a very crucial role in the world economy today. In a globalized society that we exist in, the IMF acts as monetary and fiscal guardian angel over the 185 or so member countries. The fund describes itself as “an organization of 185 countries working to foster global monetary cooperation, secure financial stability, facilitate international trade, promote high employment and sustainable economic growth, and reduce poverty.” A lofty goal indeed. To achieve this goal, the Fund is employing the best and the latest that technology has on offer. And Jonathan Palmer is overseeing that this very aspect.

Palmer is the associate director of the Technology and General Services Department of the IMF. Informally, he is the CIO at the Fund and describes himself as an ‘economist on the field’. Not such a long time ago, he was an adjunct professor at the Canberra University, when the opportunity of heading the technology division at the IMF came his way. He readily took up the challenge and seems to be enjoying his role, every bit. According to him the biggest challenge he has faced has been in terms of working with diverse and multi-cultural teams. While he admits that he is interested in open-source, he is certainly not obsessed with it and feels that proprietary standards and software also have a lot to offer. In an extended telephonic interaction from Washington DC with Shashwat Chaturvedi, Palmer spoke to Dataquest and gave an insight on what makes the IMF tick. Little surprise, it is indeed technology. Excerpts.

How does technology help the IMF meet the changing needs of member countries in an evolving world economy?
The IMF has a clear strategy for responding to the changing environment, in particular to the challenges of globalization. It’s called the Medium-Term Strategy or MTS. Technology plays a strong role in supporting this strategy. For example, since one of the requirements of the MTS is enhancing the IMF advice on financial and capital markets, we are assisting in developing new financial models for assessing risk. We are also working on reducing e-mail overload and maintaining the Fund’s efficiency, and several other aspects related to the MTS. In brief, IT helps us in meeting up with the challenge of globalization.

The IMF has been established to promote international monetary cooperation and exchange rate stability. What are the mechanisms in place to help the IMF meet its goal? How strategic is IT in meeting these objectives?
All our processes are completely enabled by IT and technology. IT has been integrated in all the core activities of the IMF, which include surveillance, financial and technical assistance. We have a broad portfolio of systems, data warehouses, publishing systems, document repositories, economic models, and a range of systems that support our Human Resources Department, our Finance Department, mission travel, etc. To say that IT is strategic will be an understatement, as it forms the basis of all our operations.

Starting with around 40 member countries in 1945 to over 180 now, how does the IMF link all these member countries and how are the statistics recorded?
The IMF has 185 member countries at present. Its organization includes Resident Representative offices in about 90 countries, three regional offices, and six technical assistance centers. All the staff in these offices are linked to the head office in Washington through our network, and we are making increased use of video conferencing, Internet telephony, instant messaging, and extranets to improve this linkage between IMF staff at headquarters and the staff in the field.

As far as collecting statistical information is concerned, we use a highly automated and secure system called the Integrated Correspondence System. We are also participating in an initiative called SDMX (Statistical Data and Metadata Exchange), which is defining formats for the exchange of data and metadata. This is relevant because to automate the collection of data and integrate data across countries, it is important to have appropriate standards. The IMF is playing a leadership role in this area.

How has the current IT infrastructure evolved over a period of time and what is the current infrastructure?
In the last year or so, we have put a lot of work into improving our infrastructure redundancy and availability. We have two data centers at headquarters, which operate independently and support each other if one of them goes out. We also have a remote DR facility, from which we can host systems in the event of a major disruption to our headquarters.

Another change is the support for new devices in the network, like Blackberries and Windows Mobile devices. We recently finished a pilot deployment of Skype and we’ll be deploying it organization-wide as a platform for instant messaging and telephony. Also, we are exploring software as a service. We are working towards keeping up-to date with the latest that technology has to offer. The two key drivers behind all our initiatives are high availability for our global workforce and cost-effectiveness, all in the framework of our MTS, as I mentioned before.

A major accusation against the IMF is the apparent lack of transparency, are there some initiatives in place to deal with this as well?
We have made enormous changes in accessibility to information and our website has been a key mechanism for that. We now publish the vast bulk of the material that we produce on surveillance, at the global, regional, and country-specific levels. The challenge for us now is to recognize that it’s not only a question of putting the information in the public domain, but communicating effectively about what the information means. That’s a challenge that technology can support. We are already doing much more video presentations through our website, we have plans to provide enhanced support for multiple languages, and right now we are doing some very interesting work on data visualization.

You have been associated with the Australian Bureau of Statistics, and the New Zealand Statistical Authority, and have been adjunct professor at the University of Canberra. So can you talk about the transition of roles? How does it feel to be a CIO of an organization that has such a major role to play in the global economy?
In many ways my role is not unlike the ones I’ve had in public sector institutions. I’m continuing to work in organizations that have a very strong public mission and do important work. But there is a difference in the global dimension this time. Personally, I’m enjoying the diversity and high caliber of the staff at the IMF.

Feature: Saving the Royal Bengal Tiger


HERE is an article of mine, recently published in the Dataquest Magazine, wherein I discuss about how technology can be employed to save big cats, especially the Royal Bengal Tiger, in India..A very special article for me…

(http://www.dqindia.com/content/egovernance/2007/107062701.asp)
——————————————————————————

The Last Roar?

Can technological innovations like GIS, GPS, Radio Telemetry save India’s national animal from certain extinction? The good news is that transformation is underway, the bad, it isn’t fast enough. Fortunately, technology might just rescue what humans have forsaken…

Wednesday, June 27, 2007

Tiger count tells a sorry tale’ screamed the headline of a leading newspaper on a sultry May morning. It was a page one lead story: a bolt from the blue for Indians who woke to the harsh reality that the National Animal had suffered a body blow. It was not the first time that such a thing had been reported. Over the past few years, there have been quite a few sting operations conducted by media houses to present the real picture. Experts have been crying hoarse over the catastrophic decimation of the species. Yet, f or the past many years, the country was lulled into a fall sense of security that all was well in the 30 odd wildlife reserves in India. The babus claimed time and again that not only were the tigers safe and sound, in fact they were thriving in the tranquility of the reserves. Nothing could be further from the truth!

The Wildlife Institute of India (WII) based in Dehradun, burst that bubble by presenting the truth to the nation last month. The preliminary report released talked of how tiger numbers had reduced quite drastically. According to WII estimates the tiger population in 16 of the country’s 28 tiger reserves in 2006-07 had fallen by over 50%. From 1,233 tigers in 2002, the reserves across Madhya Pradesh, Rajasthan, Maharashtra and Chattisgarh currently house only 490 or so tigers.

The story was picked and splashed across the media, some papers even terming it as the worst decline in tiger population since 1973. There were numerous articles, features in papers, debates and discussions on TV channels. Everyone wondered what had gone wrong, and that too so suddenly. How in the world was that possible? What had happened to all those tigers? Had they been killed for skin or bones, or had they just vanished? These were questions that were troubling every mind.

The answer was fairly simple: technology had ‘happened’ to the tigers. Or, more appropriately, to the art of counting tigers. Indeed for over two decades the science of census had been turned into an art by the officials manning the 28 tiger reserves in India. Using pugmark methodology, the census artists arrived at numbers that were vastly exaggerated and no way matched the numbers on the ground. That was the very case in Sariska.

—————————————————————-
Number of Sanctuaries and Parks for Wildlife in India
Number of national parks: 80
Total area of national parks sq km (5% of forest area): 36,882
Number of sanctuaries: 501
Total area of sanctuaries (17% of forest area): 120,052 sq km
Total protected area (22% of forest area): 156,934 sq km
*************
Source: Project Tiger

Going, Going, Gone
In December 2004, a news report highlighted the fact that there were no more tigers left in the Sariska Tiger Reserve, Rajasthan. It came as a rude shock, as till then the official record showed a healthy population of around 25-30 tigers in the reserve that is spread over an area of about 880 sq km, with three core areas (areas where no human are supposed to be present) and buffer zone (forested area with tribals).

In the wake of the uproar that followed, Prime Minister Manmohan Singh announced a CBI inquiry into the matter and even set a Tiger Task Force to study the issue and suggest remedial measures (the report was submitted and yet no action was taken, as usual).

“Sariska was simply a case of inaction. The poachers dealt a fatalistic blow to the few tigers that existed there. While all the time officials were talking about 25 tigers in Sariska. Every researcher knew it was a farce, because no one had sighted a tiger for quite many years. Those were just paper tigers, that existed solely on paper,” says Valmik Thapar, a renowned tiger conservation activist and author of quite a few books on the tiger, the latest being, The Last Tiger. He was also on the Tiger Task Force set up by the Prime Minister.

A Tale of Tall Numbers
At the turn of the twentieth century, there were an estimated 40,000 tigers in India. Indiscriminate hunting and loss of habitat resulted in a drastic dip in numbers, and by the ’70s, not more than a few thousand remained. There was international uproar as India houses close to 50% of all tigers, and if they disappear here it would effectively mean the end of the species. The future of the tiger looked dreary. The Prime Minister, at the time, Indira Gandhi, deciding to take matters in her hands set up a team to study the issue and come up with a plan to save the tigers. The team was headed by erstwhile ruler of Jammu & Kashmir, Dr Karan Singh. Based on the report submitted by the team, Project Tiger was launched amidst much fanfare in 1973.

Initially, Project Tiger was conceived for six years, from 1973-79, and was launched by setting up eight tiger reserves across the country (Sunderban was added at the launch, taking the number to nine). Over the years, the project received much success and was even termed as the ‘most successful wildlife conservation program’ in the world, by reputed international agencies. Gradually, the number of tiger reserves increased to 30 and by 2005, India had 90 national parks and 501 wildlife sanctuaries.

Thus, going by official figures, there were over 3,600 tigers in India by 2002. But the real story was something else. The figures had been arrived using pugmarks, which according to many is a faulty procedure. As it was solely based on human discretion, the figures could be easily inflated to reflect an increase in numbers while there was none.

“Pug mark census methodology was turned into an art by the officials. It was used to fabricate numbers. By the 90s, we all knew that the tiger population was declining, yet the reports would show that the numbers had actually increased,” says India’s premier tiger scientist and conservation activist Raghu Chundawat

Arriving at the Right Numbers
In the latest survey conducted by WII, the institute used advanced technology like DNA sampling and camera trap techniques to arrive at the current figure. The WII estimates showed that tiger numbers had fallen in Madhya Pradesh by 61%, Maharashtra by 57%, and Rajasthan by 40%. Compare this with the government’s first tiger census; conducted under the Project Tiger initiative, begun in 1973, it counted 1,827 tigers in the country that year. Since then the tiger population saw a steady rise to reach 3,700 tigers in 2002. Use of technology has effectively curtailed the numbers by half.

“Tiger count is a wrong terminology being used by everyone. We are not counting tigers; we are estimating the density of the tigers in a given area using scientific methodologies to arrive at the figure. Numbers by itself can be quite an erroneous game,” says BC Choudhary, senior scientist, WII.

Camera trapping is one of the technologically advanced ways of estimating the population of tigers. In this, cameras are placed across the jungle and whenever a tiger crosses a beam, the camera captures an image. As every tiger has a unique stripe coat, there is no question of the numbers being duplicated.

“Fortunately, after 40 years of wasting public money and effort, the pugmark census total count methods have been abandoned. I have been saying for 22 years that it did not work but the bureaucratic lethargy carried it so far…at great cost to tigers and to tax payers. Now even the government appears to be switching over to sampling-based methods I have used since early 1990’s,” says renowned Indian scientist, K Ulhas Karanth, director, India Program & technical director, Tigers Forever, Wildlife Conservation Society.

“Camera trapping is useful and cost effective for getting good estimates of tiger numbers in critically important reserves and populations. It is not meant for recording tiger presence across large areas or states or the country. It is a powerful and specific tool for a specific purpose,” adds Karanth.

The Poaching Menace
Without an iota of doubt, the biggest threat to the tiger is from the two-legged animal that violates the park rules and comes armed with sophisticated guns. Poaching and illegal trafficking of tiger body parts is big business. Data available at Project Tiger Directorate indicate that out of 173 deaths of tigers during 1999-2004, 83 were due to poaching. Out of the remaining, 60 deaths were due to natural causes, 13 due to electrocution, 7 due to poisoning, and 10 due to infighting. Thus, loss of tiger life due to poaching, poisoning, and electrocution works out to 103, which accounts for more than 60% of tiger deaths, reveals a CAG report for the year ended titled, “Conservation and Protection of Tigers in Tiger Reserves.”

Selling tiger skin and body parts is a highly lucrative proposition in foreign and domestic markets. Apart from tiger skin, nails, bones, tooth, fat, male genital, etc fetch anything from $300-500 per item. Tiger parts are high in demand in China, where they are used in traditional medicines. Other countries like Taiwan, Vietnam, Thailand, Singapore, are also known to be destinations for tiger parts. Not only those, sheikhs and the emirs in the Middle East are known to pay high price for tiger skin.

There are no two ways about it. If you want to save the tiger, you need to stop the poaching. Technology can be employed in the battle against poaching in a variety of ways, right from investigation to prosecution.

Sujoy Banerjee, director (Species Conservation), World Wildlife Fund (India) talks of how in the UK, the government has used a high-tech application to dissuade people who were stealing rare birds’ eggs from the nests: “The authorities identified these poachers and noted the registration number of their vehicle. Now, whenever these vehicles are found moving in such protected areas or towards Scotland (as these poachers also shifted activities to Scotland), the movement is tracked using cameras fitted by the roadside. This has helped in curbing their activities largely.”

A Black-book of Poachers
If catching a poacher was tough in India, getting him prosecuted is infinitely more tough. The big problem is that poachers are always on the move and strike at different reserves each time. This creates a problem for the police, as they are unable to corroborate the past history of the poacher in front of the magistrate, resulting in the poacher being let off easily. If there could be a central repository of all these crimes against wildlife it would be so much easier to prove cases against poachers.

“Analyzing the records in our database allows us to monitor trends in wildlife crime, which, in turn, allows us to advise enforcement officials. Secondly, we often find that poachers and traders linked to organized wildlife crime networks tend to stay in the business all their lives, and work in close knit cliques of relatives and family friends. Our database allows us to track these criminal networks as well,” she adds.Thus, if the WPSI team finds that a recently arrested individual has been arrested/convicted before for a wildlife crime, they pass the information to the public prosecutor so that the criminal’s bail application can be opposed more strongly. This database is regularly updated at WPSI and is a big asset in the fight against poaching.

Roger…Roger
Effective communication is the key to protection from forest fires, poaching, timber felling, grazing, encroachments and other illegal activities. While the poachers have been known to use sophisticated communication devices like wireless sets, UHB Radio, and even ham radio, the state of the forest guard is abysmal. “As many as nine tiger reserves- Manas, Valmiki, Indravati, Melghat, Pench (Maharashtra), Tadoba-Andhari, Periyar, Sariska, and Ranthambore were not equipped with adequate means of communication to counter illegal activities. In the Manas Tiger Reserve, four ranges, fourteen beats and two check posts were functioning without wireless network,” the CAG audit states.Not only that, even in sanctuaries where the communication network did exist, in most the devices were damaged. The Nagarjunsagar Tiger Reserve could not utilize the available communication/wireless network as per the advice of police authorities due to extremist activities in the area, and no persuasive action was taken by the forest authorities as of March 2006.

It is obvious that the Project Tiger officials do not pay much attention to communication among the guards. The consequences of such laxity are quite disastrous, and in the end it is the tiger population that takes the hit.

On the Tiger’s Trail
One of the best known methods to trail a tiger or a lion is through radio telemetry. In this, a tiger is tranquilized and a collar containing a radio antenna is put on the tiger. A researcher armed with a receiver can easily track a tiger with the aid of radio telemetry.

“Radio telemetry is a very good example of how technology can help. The tiger, as a species, is difficult to see. By nature tigers are very elusive and the quite hard to spot in the tall elephant grass. Also, it is a fact that they travel quite long distances. A male tiger often lords over an area of around 250 sq km, thus, without radio telemetry it would be hard for a researcher to track him down,” says Chundawat. In the past Chundawat has had immense success in discerning tiger behavioral patterns using radio telemetry and satellite collars. But then the red-tape stalled all Chundawat’s research efforts.”The biggest issue in use of technology, say radio telemetry or chemical immobilization, is the problems of getting research permissions. The officials managing wildlife are usually rather ignorant of science and technology and the utility of sound research. As a result, despite tremendous support to technology in other fields like medicine, ICT, agriculture etc, the field of conservation is held back by archaic mindsets. Sadly, our science policy leaders are ignoring this problem,” says Karanth.

Mapping the Forest
Another ambitious project launched by the government was digital mapping of all the wildlife sanctuaries and tiger reserves in India. It is important as an understanding of spatial factors such as connectivity between forest patches can help managers plan where and how to assign resources and efforts in general. Geographical Information Systems (GIS) are an essential tool to both wildlife managers and researchers alike.”With the development of technology, use of Geographical Positioning System (GPS) is also coming into vogue. It has been used to track the movement of Siberian cranes and sea turtles. Some tigers radio collared recently have a small GPS attachment which puts itself on at a certain time of the day as programmed, and then searches for satellites for fixing the location of the animal. The unit switches off by itself. It also transmits data to handheld devices which can be transferred onto a GIS,” says Banerjee.
Yet the signs have not been quite encouraging on the mapping front. For instance, in March 2004, the Ministry of Environment and Forests sanctioned a project at the cost of Rs 1.39 crore for mapping wildlife sanctuaries and parks by the WII.

The project was to be completed in 36 months with the aim to generate accurate, reliable and latest base line spatial information on forest types and density (using satellite imagery) and topographic features (supplemented by latest satellite imagery), On the completion of the pilot, WII was to take up the complete mapping and monitoring of wildlife sanctuaries in coordination with the Indian Institute of Remote Sensing (IIRS), National Remote Sensing Agency (NRSA) and Aligarh Muslim University. But according to the facts in the CAG report, “out of the five sites selected for this pilot project, three were tiger reserves, namely Corbett, Tadoba-Andhari and Dudhwa. The identified targets included generation of satellite data by July 2004 and securing the Survey of India (SoI) topographic maps by September 2004. However, both activities were not completed even as of March 2006. As against the projected expenditure of Rs 1.20 crore in the first two years, Rs 0.73 crore was released of which only Rs 0.30 crore was spent till the end of February 2006, indicating poor progress of the project.”

“Project tiger creates much hype about use of technology, but the performance on ground has been rather poor. They claimed to provide web-based access to their survey data three years ago. I see no sign of it at all even today. A lot of hype really,” says Karanth.

George Schaller, a globally renowned field biologist, and known as the greatest naturalist of the 20th century, sums up the issue beautifully. “Field biologists, such as Karanth and Chundawat, can use technology in the form of satellite radio-collars, camera-traps, DNA analysis of scats and other techniques to determine population size, movement patterns, and other aspects. That provides extremely valuable information. Such knowledge is essential for conservation-but it is not conservation. Conservation, in the final analysis, is culture, economics-and politics,” he mentions.

Baghbahadur and More
Indeed what Schaller says is quite true. In the end, conservation is also politics and the only way (read the right way) to influence politicians is through public opinion. If there is a huge rabble raised, usually the politician will have to back down. Internet can play a big role in influencing the opinion by projecting the true state of affairs to millions of individuals across the globe. Quite many tiger activist have woken up to the potential of the net.Ajay Sartape is one such individual. He has recently launched a tiger portal, http://www.baghbahadur.com/, in conjunction with a friend. He aims to consolidate all the disparate information on the internet on a single site. “Not only do we intend to provide information but also attempt to promote conservation. The problem with tiger conservation is that the common person is unable to associate him or herself with the cause. By providing different means and channels, we hope to spread the word on tiger all across India. The success of Baghbahadur will be in influencing policy decisions through public opinion,” says Sartape. Tall aim, but Internet has been known to connect individuals like no other medium did so in the past. Thus it could play a big role in conservation. “Internet has revolutionized how information on conservation issues is disseminated. It also improves fund raising opportunities for conservation groups and the dissemination of knowledge about the species in question,” opines Banerjee.

Any Silver Lining?
The complete report by WII will be out by December 2007. In such a gloomy scenario, one wonders if there is really any hope left for the lord of the jungle. “Technology can enable informed decision in a timely fashion and more importantly make the whole process transparent, making people accountable for their decisions and actions. This is what has been missing in Project Tiger,” says Chundawat.Thapar sounds skeptical. “The real fact of life is that tiger protection is not a priority for the government. So all it does is provide lip service. In many ways, it was only under the Gandhi family’s direct rule that the tiger was safe. After Rajiv Gandhi’s death, the numbers have been fast declining. Even the current government does not seem keen. There was the talk of setting up a state-of-the-art Wildlife Crime Bureau, but it exists only on paper,” says Thapar, ending with a note of caution, “There will not be any recovery. Damage done is damage permanent.”Meanwhile, Rajesh Gopal, director, Project Tiger was away in Europe and Ganga Singh, additional director, India Ecodevelopment Project, Ministry of Environment and Forests, would not comment, “as we need clearance from the ministry,” he said.

But can technology come up with a solution that will solve the issue, something like what was shown in Jurassic Park. Adam Barlow from the Sunderbans Tiger Project in Bangladesh dismisses the thought. “Blind faith; this is generated with respect to the idea that technological advances can combat large scale long term environmental damage-not with current knowledge, which has been proven insufficient, but with the hope that ‘well we’re bound to think of something!’ The reliance on advancement in technology to solve social and environmental problems is a fallacy, made dangerous by growing acceptance,” he says.As Schaller says, with characteristic brevity and clarity, “India has to decide whether it wants to keep the tiger or not. It has to decide if it is worthwhile to keep its National Symbol, its icon, representing wildlife. It has to decide if it wants to keep its natural heritage for future generations, a heritage more important than the cultural one, whether we speak of its temples, the Taj Mahal, or others, because once destroyed it cannot be replaced. If the answer is yes, then plans can be made and implemented.”

Jim Corbett, the famous hunter and conservationist, had sounded a warning more than half a century back in his famous book, the Man-eaters of Kumaon. “I am convinced that all sportsmen, no matter whether their viewpoint has been a platform on a tree, the back of an elephant or their own feet, will agree with me, and that is, that a tiger is a large-hearted gentleman with boundless courage and that when he is exterminated, as exterminated he will be unless public opinion rallies to his support, India will be the poorer by having lost the finest of her fauna.” So let us not bat an eyelid and rally to the tiger’s support, there might just be a chance, a rather slim one.

***x***


Chip Inside

In 2002, the leopards in Maharashtra had turned into public enemy number one. There were numerous cases of leopard attacks. Wildlife biologist and vet, Aniruddha Belsare, decided to study the human-animal conflict. He tagged a few leopards that were caught in the proximity of human settlements, and later released in the wild. The chips implanted were passive alphanumeric tags. “The study threw up some very interesting clues. As these leopards were released in the wild, in certain cases close to 400 km from their home territory, they were disturbed. Because of that they started attacking cattle, animals and even humans. They became aggressive due to habitat displacement,” he said. Belasare also talks how leopards try and hone into their home territory, in one case a leopard that was chipped was found over a 100 kms away from the location he was released, he was trying to return to his original base. “Technology can be used greatly to study conflict in animals with radio telemetry and embedded chip,” says Belasare.

DNA Profiling
As it is hard to identify a tiger due to its elusive nature, researchers are all the time trying to find a solution. Non-invasive DNA-based techniques, which identify individual species and animals with the molecular analysis of collected hair and scat samples. With the profiling in place, trends in tiger behavior will be easier to discern. There are two institutes in India that are currently working on DNA profiling, WII and the Centre for Cellular and Molecular Biology in Hyderabad. “DNA analysis on wild population can also help monitor poaching trends. For instance, if we had a DNA profile of all the tigers, and if we caught a tiger skin, we could compare it to the database and find out where this tiger had been killed. If there are more cases from a certain forest reserve or reserves, you know where these poachers are active,” says Chundawat.

Tracking Elephants with GPS
WWF in India is using GPS to track elephants in Assam. “We are tracking the movement of elephants in Assam (by following elephant herds, or using secondary indicators of elephant movement such as dung, footprints, etc) by using GPS. The data is transferred on to a GIS, which gives us critical corridors”. says Banerjee.The study has thrown up a lot of data on animal behavior and helped understand human-animal conflict better.

By using this technology, we also come to know more about their “errant” behavior, for eg, the very recent storming of Hojai town in Assam due to which a curfew had to be imposed for two days,” he adds.The Wildlife Protection Society of India (WPSI) has created one such database, a veritable of who’s who of Indian poachers. “At WPSI, we work extensively with enforcement agencies such as the Forest Department and the police. We have found that one of our most effective tools is our wildlife crime database which has over 15,000 records of wildlife crime from in and around India,” shares Onkuri Majumdar, senior project officer and the legal program, WPSI.

Electronic surveillance in the jungle
Wildland Security, a U.S. based company, has designed electronic surveillance technology for monitoring of protected forests. Labeled as TrailGuard, the remote electronic surveillance technology increases the effectiveness of protection efforts both in terms of the cost effectiveness of monitoring as well as by increasing apprehension before poaching occurs. TrailGuards work by detecting human intruders as distinct from other animals that may be in the forest. The principle discriminate that we use are magnetometers (metal detectors) buried near the principle trails poachers are likely to use. Humans would be the only ones carrying iron objects-machettes and guns, for example. The detection events are then transmitted by a radio data link to a low-cost satellite internet gateway located in the canopy. And hence to the internet and a web server where the event is recorded and then emails, GSM SMS text messages and satellite pages are sent to a designated list of parties such as the chief ranger and, possibly, funding agencies. So within a matter of a few minutes after an intrusion is detected, patrols can be dispatched to where the event occurred, says Steve Gullick from Wildland Security.Is it costly? “The system is in a prototype stage so I can not give exact figures as to the cost once it is in full production. I think it would be in the order of $1,000 per trail site protected. I also think that in many areas there may be only 6-10 principle established trails into an area so a large number of TrailGuards would not be required to provide a greatly increased 24/7 surveillance,” says Gullick.He has been approached by Biswajit Mohanty of Wildlife Society of Orissa for elephant protection and Ulhas Karanth as well. Probably TrailGuards can solve the issue of monitoring the parks, a big issue in India.

—EOM