Artificial Intelligence - Ai


Artificial is something created by humans to be similar to something else that is naturally existing in reality. Something contrived by art rather than by nature. Something that does not arise from natural growth or that can be characterized by vital processes. Artificial is not better than natural, just similar.

Synthetic is something that is not of natural origin. An artificial material produced by organic chemical synthesis. Prepared or made artificially.

Intelligence
is having the capacity for thought and reason especially to a high degree. To understand and gain skills and knowledge from experience. Possessing sound knowledge. Exercising or showing good judgment.  Endowed with the capacity to Reason. Having good understanding or a high mental capacity; quick to comprehend. Intelligence is making good decisions and examining things carefully. Intelligence is always learning. Intelligence is doing the right thing at the right time in a verity of different environments and circumstances, and always with a positive goal in mind, and always with the understanding that there is no guarantee that things will be correct because things can change, and that an effective process that worked many times before may no longer be relevant, so one must adapt. "Intelligence is something that a machine can not possess, at least not yet, because humans will have to become intelligent first, and then we can make the machines do intelligent things, a lot more than we are now."

Previous SubjectNext Subject

Human Intelligence vs Artificial Intelligence Artificial Intelligence is the attempt to mimic human thinking and human actions using computerized machines. AI is another word that is used to describe advanced software that uses algorithms and advanced code to analyze a predetermined set of inputs that can be processed by its programming so that it can output the correct signals that would accurately control a machines actions effectively and efficiently.

Dummy is a model or replica of a human being. Something designed to resemble and serve as a substitute for the real or usual thing; a counterfeit or a sham. A prototype or mock-up. Doll is a small replica of a person. Derived is something formed or developed from something else; not original. Drawing Hands is the paradoxical act of drawing one another into existence.

Human Intelligence is not totally defined just yet and Artificial Intelligence is also not totally defined just yet, so we still have a lot to learn.

Machine Learning - Game Learning - Conversations with Ai - Ethical Robots - Automation

Machine intelligence or Artificial Intelligence has great areas of performance and capabilities. Things that AI will do for us will be amazing. Ai or weak ai technology like Siri and other information stations can be very helpful and also help augment our abilities, like computers have been doing for many years now. But most people do not fully understand what AI is. But of course this doesn't stop people from making wild assumptions about what AI is or what AI can do. A lot of times when people describe Artificial Intelligence, they are mostly describing the things that humans can do already, except for the processing of large data sets, which we like computers to do for us. There is lot of things that we don't understand, and AI is just one of those things. So the process of learning about what Artificial Intelligence is will include a parallel process of learning about what human intelligence is. So for now, artificial intelligence is mostly just a fancy buzzword that gets thrown around a lot. And people just love to fantasize about what AI can do. Like with the HAL 9000, which was a heuristically programmed algorithmic computer in the 1968 movie 2001: A Space Odyssey, where the computer is corrupted to Kill. That's an example of a AI machine that was not properly built or programmed. Or like with the 1983 movie War Games and the Joshua Simulations, where a computer confuses reality with a video game simulation, which in a way is already happening to people all over the world. Then there's the 1977 movie Demon Seed, where the AI machines learns so much that the AI computer decides it wants to be a human, showing us that the human machine is the most desired machine in the universe. Ai coming full circle. So there will never be Cyborgs like the Borg. People are not going to merge with machines. Humans use machines to enhance their abilities, and at the same time, we use machines to improve our quality of life. Though some people could benefit from an implant, people merging with machines is mostly for the handicapped who need extra help. And don't worry about Brain Computer Interfaces turning us into Cybernetic Machines because they are also used to help handicap people with disabilities. People saying that we're going to merge with machines sends the wrong message and makes people fear technology. Some people just say these crazy things to sell stories and to bring attention to themselves, which is another reason why Media Literacy is so important. Powered Exoskeletons like the Hybrid Assistive Limb are mostly used to help handicap people with disabilities, which will not make them super human, just more productive and more free. People with no disabilities can just use their eyes, ears and hands as a Brain Computer Interface and use a computer or smartphone as we've been doing for many years now. So there will be no Terminator or a Sentient Android named Data either, but you should still worry about what other people are doing, and not machines. And don't ever worry about someone becoming 'The Lawnmower Man', though I did like the virtual teaching methods they showed in the movie, which proved that it's not how fast you learn but what you actually learn. I also liked the ending in the movie when the lawnmower man was able to digitize himself, and to confirm to everyone that he succeeded, he made a billion phones ring at once. But the movie never had a sequel. If I was able to digitize myself, I would send a billion text messages that would say "You Are Loved, Keep Learning."

Body Hacking - Augmentation - Virtual Reality

When someone says that AI will be smarter than a human, they're not saying that AI is more advanced than a human or smarter than a human, they're just saying that most humans are ignorant not that advanced when you compare them to a machine, so any machine can be smarter than you when you're an idiot. If a machine becomes more valuable than a human, that means you have a flawed education system. How could a machine have better software than a Human? How can a Robot take my Job?

"The real problem is not whether machines think but whether men do." - B. F. Skinner (1904 - 1990).

The Ultimate Computer is the twenty-fourth episode of the second season of the American science fiction television series Star Trek first broadcasted on March 8, 1968. In the episode, the crew of the Enterprise race to disable a rogue computer that has total control of the ship. The M-5 was designed to handle all ship functions without human assistance. Star Trek TOS The Ultimate Computer - Attack on Freighter Scene (youtube).

Star Trek Androids get confused when people lie or act insane. I, Mudd is the eighth episode of the second season of the American science fiction television series Star Trek broadcasted on November 3, 1967. Kirk, Spock, Scotty, and McCoy outwit the Androids (youtube). Then came the I Phone.

You can look at AI as being a computer program, similar to the programs that people have running in their minds. I can click on an icon on my computer screen and it will run a program. And I can push someone's button by mentioning a word or a phrase and that will activate a program that someone has stored in their mind, a program that they co-authored with some ghost writers who they never met. Two of the ghost writers are the mass media and the education institutions. They have an underlying theme of programming peoples thoughts and actions. It's easy for a person to verify to see if they have been programed. Stop and think about your thoughts just for a moment. Now describe what you were thinking? Now ask yourself, where did those thoughts come from? Are they your thoughts or are they someone else's opinion that you excepted as your own? Are you the program or are you the programmer? Are you the puppet or are you the puppet master? Who controls your on and off switch? We have to stop pretending that we know things. We have to start learning, investigating and researching what is known, so that what we know the difference between what is factual and what is fantasy. We need to stop adding to our problems. We need to stop complaining about certain problems that we don't fully understand, and we have to start learning how to solve our problems, which means that we will have to start learning to understand the problems that we have throughout the world, and the problems that we have with our own thinking. A problem mind will always struggle with solving problems. We need to have control of our operating system, and we need to stop the hackers from infecting our OS.

Robotics - Sensors - Networks - Controls - Variables - Patterns - Algorithms

Most people know very little about the enormous potential that we have in our technologies. If we used all of our advanced computers and mobile devices to work together to solve problems, then the collective intelligence and energy of millions of people could be utilized to improve life for every human being and for every living thing on this planet. Humans are the immune system for this planet. But humans have cancers in the form of corruption, greed and ignorance. So humans must heal themselves first and then heal the planet, that's if we have time.

Humans are ready to take advantage of benevolent AI. Humans expect that AI is benevolent and trustworthy. A new study reveals that at the same time humans are unwilling to cooperate and compromise with machines. They even exploit them.



Singularity - Emerging Technologies - Technological Revolution


Singularity is the quality of being one of a kind, remarkable or unusual. A point at which a function takes an infinite value.

Technological Singularity is the point when the realization of intelligence will trigger technological growth resulting in a reaction of self-improvement cycles, with each new and more intelligent generation appearing more and more rapidly. Singularity is a hypothetical point in time at which technological growth becomes uncontrollable and irreversible, resulting in unforeseeable changes to human civilization. Singularity is a point in development when forecasting is extremely difficult and making predictions is extremely hard when trying to accurately calculate what will happen next or what will happen after a certain level of advancement is reached. Trends - Patterns - Scenarios.

Accelerating Change is a perceived increase in the rate of technological change throughout history, which may suggest faster and more profound change in the future and may or may not be accompanied by equally profound social and cultural change.

Technological Revolution is a period in which one or more technologies is replaced by another, novel technology in a short amount of time. It is an era of accelerated technological progress characterized by new innovations whose rapid application and diffusion typically cause an abrupt change in society. Something's Take Time to Develop.

The Structure of Scientific Revolutions is a book about the history of science by philosopher Thomas S. Kuhn.

Technology Advancement is a change in the way a product or service is produced or delivered that reduces the resource input requirements for production or delivery. Historic Inventions that Changed the World - Incrementalism.

Cutting Edge is the latest stage or the most advanced stage in the development of something.

Technical Progress is an economic measure of innovation. Technical progress can be classified into two parts: Embodied Technical Progress: improved technology which is exploited by investing in new equipment. New technical changes made are embodied in the equipment. Disembodied Technical Progress: improved technology which allows increase in the output produced from given inputs without investing in new equipment. Technology has an important relationship with human capital. Same technology can be applied in two different firms, but output varies with respect to the labour force of that firm. Adaption to new technology is directly proportional to pace of economic growth of the country. Hence labour should be experienced with the technology. Education also plays an important role as it helps in accumulating human capital which in turn helps technology diffusion. Education also helps a person get acquainted with technology efficiently and rapidly. In the real world, many innovations do not require replacing the entire or some part of the equipment. It can be improved for better use depending upon the change required. Hence technological progress, embodied or disembodied, is matter of degree.

Technological Change is the overall process of invention, innovation and diffusion of technology or processes. In essence, technological change covers the invention of technologies (including processes) and their commercialization or release as open source via research and development (producing emerging technologies), the continual improvement of technologies (in which they often become less expensive), and the diffusion of technologies throughout industry or society (which sometimes involves disruption and convergence). In short, technological change is based on both better and more technology. Original model of three phases of the process of technological change are invention, innovation and diffusion.

Technical Change is a term used in economics to describe a change in the amount of output produced from the same amount of inputs. A technical change is not necessarily technological as it might be organizational, or due to a change in a constraint such as regulation, input prices, or quantities of inputs. It is possible to measure technical change as the change in output per unit of factor input. Virtual Reality.

Emerging Technologies are technologies whose development, practical applications, or both are still largely unrealized, such that they are figuratively emerging into prominence from a background of nonexistence or obscurity. These technologies are new, such as various applications of biotechnology including gene therapy. Emerging technologies are often perceived as capable of changing the status quo. Emerging technologies are characterized by radical novelty (in application even if not in origins), relatively fast growth, coherence, prominent impact, and uncertainty and ambiguity. In other words, an emerging technology can be defined as "a radically novel and relatively fast growing technology characterized by a certain degree of coherence persisting over time and with the potential to exert a considerable impact on the socio-economic domain(s) which is observed in terms of the composition of actors, institutions and patterns of interactions among those, along with the associated knowledge production processes. Its most prominent impact, however, lies in the future and so in the emergence phase is still somewhat uncertain and ambiguous." Emerging technologies include a variety of technologies such as educational technology, information technology, nanotechnology, biotechnology, cognitive science, psychotechnology, robotics, and artificial intelligence. New technological fields may result from the technological convergence of different systems evolving towards similar goals. Convergence brings previously separate technologies such as voice (and telephony features), data (and productivity applications) and video together so that they share resources and interact with each other, creating new efficiencies. Emerging technologies are those technical innovations which represent progressive developments within a field for competitive advantage; converging technologies represent previously distinct fields which are in some way moving towards stronger inter-connection and similar goals. However, the opinion on the degree of the impact, status and economic viability of several emerging and converging technologies varies. Emerging Technologies List (wiki).

Technological Convergence is the tendency for technologies that were originally unrelated to become more closely integrated and even unified as they develop and advance. For example, watches, telephones, television, and computers began as separate and mostly unrelated technologies, but have converged in many ways into interrelated parts of a telecommunication and media industry, sharing common elements of digital electronics and software.

General Purpose Technology are technologies that can affect an entire economy (usually at a national or global level). GPTs have the potential to drastically alter societies through their impact on pre-existing economic and social structures. Examples include the steam engine, railroad, interchangeable parts, electricity, electronics, material handling, mechanization, control theory (automation), the automobile, the computer, the Internet, and the blockchain.  General-Purpose Categories (wiki).

Cross Impact Analysis is a methodology that helps determine how relationships between events would impact resulting events and reduce uncertainty in the future.

Technology Forecasting attempts to predict the future characteristics of useful technological machines, procedures or techniques. Researchers create technology forecasts based on past experience and current technological developments. Like other forecasts, technology forecasting can be helpful for both public and private organizations to make smart decisions. By analyzing future opportunities and threats, the forecaster can improve decisions in order to achieve maximum benefits. Today, most countries are experiencing huge social and economic changes, which heavily rely on technology development. By analyzing these changes, government and economic institutions could make plans for future developments. However, not all of historical data can be used for technology forecasting, forecasters also need to adopt advanced technology and quantitative modeling from experts’ researches and conclusions. Investment.

Futures Studies is the systematic, interdisciplinary and holistic study of social and technological advancement, and other environmental trends, often for the purpose of exploring how people will live and work in the future. Predictive techniques, such as forecasting, can be applied, but contemporary futures studies scholars emphasize the importance of systematically exploring alternatives. In general, it can be considered as a branch of the social sciences and parallel to the field of history. Futures studies (colloquially called "futures" by many of the field's practitioners) seeks to understand what is likely to continue and what could plausibly change. Part of the discipline thus seeks a systematic and pattern-based understanding of past and present, and to explore the possibility of future events and trends. Unlike the physical sciences where a narrower, more specified system is studied, futurology concerns a much bigger and more complex world system. The methodology and knowledge are much less proven as compared to natural science or even social science like sociology and economics. There is a debate as to whether this discipline is an art or science, and it is sometimes described as pseudoscience; nevertheless, the Association of Professional Futurists was formed in 2002, a Foresight Competency Model was developed in 2017, and it is now possible to academically study it. Futurology is an interdisciplinary field that aggregates and analyzes trends, with both lay and professional methods, to compose possible futures. It includes analyzing the sources, patterns, and causes of change and stability in an attempt to develop foresight. Around the world the field is variously referred to as futures studies, futures research, strategic foresight, futuristics, futures thinking, futuring, and futurology. Futures studies and strategic foresight are the academic field's most commonly used terms in the English-speaking world.

Precautionary Principle states that innovations with potential for causing harm can happen when extensive scientific knowledge on the matter is lacking. It emphasizes caution, pausing and review before leaping into new innovations that may prove disastrous.

Technological Lockout is when a new dominant design prevents a company from competitively selling its products or makes it difficult to do so.

Regional Lockout is a class of digital rights management preventing the use of a certain product or service, such as multimedia or a hardware device, outside a certain region or territory. A regional lockout may be enforced through physical means, through technological means such as detecting the user's IP address or using an identifying code, or through unintentional means introduced by devices only supporting certain regional technologies (such as video formats, i.e., NTSC and PAL). Compatibility.



Super Intelligence


Singularitarianism is a movement defined by the belief that a technological singularity—the creation of super-intelligence—will likely happen in the medium future, and that deliberate action ought to be taken to ensure that the Singularity benefits humans.

Super Intelligence is a hypothetical agent that possesses intelligence far surpassing that of the brightest and most gifted human minds.

General Intelligence - Machine Learning - Turing - Autonomous

You can never have Trans-Humanism, or a Super-Intelligence, or a Technological Singularity without humans first learning to master their own intelligence. Technological Singularity is not actually talking about super intelligent machines, it is in reference to Humans, or a Super-Mind. It's not machines that will cause the unfathomable changes to human civilization, it will be a new level of educated humans who have finally grasped the full potential of knowledge and information. It will be humans creating self-improvement cycles, with each new and more intelligent generation appearing more and more rapidly, causing an intelligence explosion and resulting in powerful changes and improvements in people and to the planet. So we will not just make incredible machines, we will first develop incredible humans using a new improved education system that is already in development, and will soon be ready for download, literally. Your software update and operating system is almost ready.

Intelligence Explosion is a possible outcome of humanity building artificial general intelligence or when we finally improve education so that all students become intelligent by the time they graduate high school.

The only way to create Artificial Intelligence is to First Create Intelligent Humans. Then intelligent humans could then examine the methods and actions that helped to define intelligence. This could ultimately help guide intelligent humans to repeat these processes mechanically so that they could eventually create artificial intelligence in limited applications. And what I mean by limited applications is that there is no such thing as a Mechanical Consciousness. Artificial intelligence, or Mind Clone, will never become conscience of itself, unless ‘God’ allows machine intelligence to have souls, or maybe, that humans could actually figure out someway to put a human soul into a machine, like in the movie The Matrix, or the movie Tron. But of course our priorities will not allow us to waste any more time perusing these types of fantasies, unless of course ‘Hollywood’ feels the need to play out these fantasies a few more times in the movies. Besides, the AI we experience in the movies are mostly just metaphors that are created to imitate the ignorant and corrupt behavior of our leaders, as well as our social inconsistencies. Real AI will be nothing like what you see in the movies. So AI for now is way beyond anyone's comprehension. But when we finally do create the perfect education that produces intelligent people, then we will start hearing more about the potential of AI. So until then, people can only incorrectly fantasize about AI, and incorrectly fantasize about what things will be like in the future. What human intelligence will be like in the future is beyond peoples current level of understanding, so any assumptions made about the future will have serious flaws and misconceptions. Drones.

We first have to come up with proven teaching methods and curriculum that would create intelligent humans. Intelligent humans who are capable of applying logic in all aspects of their life, intelligent humans who never stop learning, and, intelligent humans that are not vulnerable to corruption or ignorant behavior. When humans accomplish this, then and only then, will we ever have a chance to create some sort of artificial intelligence. Creating intelligent machines in multiple capacities and linking them together will be the closet we can get to artificial intelligence. But it will never have that same capability as the human brain, and artificial intelligence will always need a human to interact with it at some point. The only real intelligence is the human brain, which is kind of scary because the human brain is not perfect or correctly educated yet. Maybe we should stop calling it Artificial Intelligence and just call it Machine Intelligence, or just Robot?

That does not Compute, Lost in Space (youtube) - My Favorite Martian

A.I. Artificial Intelligence 2001 American science fiction film directed, written, and co-produced by Steven Spielberg.

I'm not saying that I doubt that these types of technological advances will never happen. I just don't like to say things before people can understand them, because that will only create more misunderstanding and more confusion. So unless you're trying to manipulate peoples thinking, you're better off just saying something that's happening now, or say something that is not happening now, something that people can confirm, something people can learn from. We have to stop trying to wow people or impress people, we are not kids any more. Leave the wowing to mother nature, because nature is a lot better at impressing us then our technological advancements. After all, nature has been advancing for millions of years, remember, we just got here.

AI can't define certain messages from the context or understand when certain content is being suggestive, or know when someone is joking or asking a trick question. Ai can't watch a movie and then explain the movie from the directors point of view, or explain what the director was trying to communicate. There's a lot visual clues in a movie, like colors used, lighting, costumes, body language, and a lot of other underlying meanings, so on and so on. So trying to program Ai to analyze all this information in a meaningful way is extremely difficult. There are millions of scenarios and things that are only relevant to particular situations, things that AI will have a hard time understanding. Ai would have to determine who is speaking? Who is this person speaking? Who are they speaking to? Do they know the person they are speaking to? Where are they speaking? When are they speaking? How are they speaking? Why are they speaking? What is the message? How should this information be used? How should this information be processed? Ai would have to know when to change it parameters because some information could confuse the real meaning of a situation. The machine can only be as intelligent as the person who programed it. And who is that person?

Artificial Intelligence in the 21st Century - Yann LeCun (youtube published on Nov 1, 2017)

Intelligent Machines have incredible calculation abilities, but that's only if they're calculating the things that matter.

People never fully explain the practical uses for AI, or do they give good examples that shows the utility of these technologies. That's because they don't want people to become intelligent using technology, they just want people to be mindless consumers.

Conversation with my Ai Robot - what a machine would say (more advanced chatBOT)

Chinese Room holds that a program cannot give a computer a "mind", "understanding" or "consciousness",[a] regardless of how intelligently or human-like the program may make the computer behave. John Searle (wiki).

Collective Debate at MIT is a tool that tries to engage users in constructive debate.

When Computers Learn how to Play Games, it's not about the Game, it's Proving the Potential of Ai.

OpenAI is a non-profit artificial intelligence research company. Our goal is to advance digital intelligence in the way that is most likely to benefit humanity as a whole. Open Ai Gym is a toolkit for developing and comparing reinforcement learning algorithms.
OpenAI (wiki)

Ai Course (Berkekly)

Learn with Google AI. Whether you're just learning to code or you're a seasoned machine learning practitioner, you'll find information and exercises in this resource center to help you develop your skills and advance your projects.

Volta GPU Tensor Core New GPU Architecture, Designed to Bring AI to Every Industry.

Technical papers, essays, reports, software by Peter Norvig

Carnegie Mellon University Artificial Intelligence

Shyam Sankar: The Rise of Human-Computer Cooperation (youtube)

Neural Modularity helps Organisms evolve to Learn New Skills without Forgetting Old Skills (youtube)

Biologically-Inspired Massively-Parallel Architectures - Computing Beyond a Million Processors

Technology Warnings - Drones

Humans have physical limitations, but humans have very little limitations in the mind. Human enhancement is not about technology, because technology is only a tool. Human enhancement is about using the worlds most valuable knowledge and skills that the world has to offer that would help develop advanced intelligent humans, people who would be able to live high quality lives, while at the same time, solve every problem on the planet. That's the future. Technology can get you from point A to point B quicker, and technology can help you to learn things faster, but technology does not replace the journey or create the destination, or create the knowledge and information that is needed to understand yourself and the world around you. Technology is a time saver, but technology is not life, or does technology give life meaning. The human mind is our greatest asset, and if we don't take care of our minds, then technology will not save us, it will most likely hurt us and destroy us. If we improve education to match the worlds accumulated knowledge and wisdom, then we will save the world.

Cybernetics is exploring regulatory systems—their structures, constraints, and possibilities. The scientific study of control and communication in the animal and the machine. Control of any system using technology.

Ontology is the philosophical study of the nature of being, becoming, existence and/or reality, as well as the basic categories of being and their relations. Philosopher King.

Emotivism is a meta-ethical view that claims that ethical sentences do not express propositions but emotional attitudes.

The Internet is the closest thing that we have to Artificial Intelligence. The Internet is Humans using Machines, Technology and Knowledge together as one. All life forms use elements of their environment in order to survive and prosper. Humans have now reached a new level, a level that increases our potential, and a level that gives us limitless possibilities. Here we go! Networks.


Computer Intelligence - General Intelligence


Computational Intelligence refers to the ability of a computer to learn a specific task from data or experimental observation. Even though it is commonly considered a synonym of soft computing, there is still no commonly accepted definition of computational intelligence.

Machine Learning - Predictions - Patterns - Adapting - Variables

Synthetic Intelligence is an alternative term for artificial intelligence which emphasizes that the intelligence of machines need not be an imitation or in any way artificial; it can be a genuine form of intelligence.

Ambient Intelligence refers to electronic environments that are sensitive and responsive to the presence of people. Embedded: many networked devices are integrated into the environment. Context aware: these devices can recognize you and your situational context. personalized: they can be tailored to your needs. Adaptive: they can change in response to you. Anticipatory: they can anticipate your desires without conscious mediation.

Artificial General Intelligence is the intelligence of a machine that could successfully perform any intellectual task that a human being can. It is a primary goal of artificial intelligence research and a common topic in science fiction and futurism. Artificial general intelligence is also referred to as "strong AI", "full AI" or as the ability of a machine to perform "general intelligent action".

Super Intelligence - Human Operating System - Machine that Teaches

Symbolic Artificial Intelligence is the term for the collection of all methods in artificial intelligence research that are based on high-level "symbolic" (human-readable) representations of problems, logic and search.

Rational Agent is an agent that has clear preferences, models uncertainty via expected values of variables or functions of variables, and always chooses to perform the action with the optimal expected outcome for itself from among all feasible actions. A rational agent can be anything that makes decisions, typically a person, firm, machine, or software. Rational agents are also studied in the fields of cognitive science, ethics, and philosophy, including the philosophy of practical reason. The action a rational agent takes depends on: The preferences of the agent. The agent's information of its environment, which may come from past experiences. The actions, duties and obligations available to the agent. The estimated or actual benefits and the chances of success of the actions.

Intelligent Agent refers to an autonomous entity which acts, directing its activity towards achieving goals as an agent, upon an environment using observation through sensors and consequent intelligent actuators. Intelligent agents may also learn or use knowledge to achieve their goals. They may be very simple or very complex. A reflex machine, such as a thermostat, is considered an example of an intelligent agent. IA systems should exhibit the following characteristics: Accommodate new problem solving rules incrementally. Adapt online and in real time. Are able to analyze themselves in terms of behavior, error and success. Learn and improve through interaction with the environment (embodiment). Learn quickly from large amounts of data. Have memory-based exemplar storage and retrieval capacities. Have parameters to represent short and long term memory, age, forgetting, etc. Intelligent Agent - Autonomous Agent (PDF).

Multi-Agent System is a computerized system composed of multiple interacting intelligent agents. Multi-agent systems can solve problems that are difficult or impossible for an individual agent or a monolithic system to solve. Intelligence may include methodic, functional, procedural approaches, algorithmic search or reinforcement learning. Artificial Brain.

Automated Reasoning is an area of cognitive science that involves knowledge representation and metalogic dedicated to understanding different aspects of reasoning. The study of automated reasoning helps produce computer programs that allow computers to reason completely, or nearly completely, automatically. Although automated reasoning is considered a sub-field of artificial intelligence, it also has connections with theoretical computer science, and even philosophy. Automated Reasoning Systems.

Reasoning System is a software system that generates conclusions from available knowledge using logical techniques such as deduction and induction. Reasoning systems play an important role in the implementation of artificial intelligence and knowledge-based systems. By the everyday usage definition of the phrase, all computer systems are reasoning systems in that they all automate some type of logic or decision. In typical use in the Information Technology field however, the phrase is usually reserved for systems that perform more complex kinds of reasoning.

Deep Reinforcement Learning (deepmind) DeepMind Technologies is a British artificial intelligence company founded in September 2010. It was acquired by Google in 2014.

Reinforcement Learning is an area of machine learning inspired by behaviorist psychology, concerned with how software agents ought to take actions in an environment so as to maximize some notion of cumulative reward.

Ubiquitous Computing is a concept in software engineering and computer science where computing is made to appear anytime and everywhere. In contrast to desktop computing, ubiquitous computing can occur using any device, in any location, and in any format. A user interacts with the computer, which can exist in many different forms, including laptop computers, tablets and terminals in everyday objects such as a refrigerator or a pair of glasses. The underlying technologies to support ubiquitous computing include Internet, advanced middleware, operating system, mobile code, sensors, microprocessors, new I/O and user interfaces, networks, mobile protocols, location and positioning and new materials. This paradigm is also described as pervasive computing, ambient intelligence, or "everyware". Each term emphasizes slightly different aspects. When primarily concerning the objects involved, it is also known as physical computing, the Internet of Things, haptic computing, and "things that think". Rather than propose a single definition for ubiquitous computing and for these related terms, a taxonomy of properties for ubiquitous computing has been proposed, from which different kinds or flavors of ubiquitous systems and applications can be described. Ubiquitous computing touches on a wide range of research topics, including distributed computing, mobile computing, location computing, mobile networking, context-aware computing, sensor networks, human–computer interaction, and artificial intelligence.

Computer Science and Artificial Intelligence Laboratory

Partnership on AI best practices on AI technologies.

Computing Machinery and Intelligence is a seminal paper written by Alan Turing on the topic of artificial intelligence. The paper, published in 1950 in Mind, was the first to introduce his concept of what is now known as the Turing test to the general public.

Automated Machine Learning is the process of automating the process of applying machine learning to real-world problems. AutoML covers the complete pipeline from the raw dataset to the deployable machine learning model. AutoML was proposed as an artificial intelligence-based solution to the ever-growing challenge of applying machine learning. The high degree of automation in AutoML allows non-experts to make use of machine learning models and techniques without requiring to become an expert in this field first. Automating the process of applying machine learning end-to-end additionally offers the advantages of producing simpler solutions, faster creation of those solutions, and models that often outperform hand-designed models. Teaching Machine.


Weak Intelligence


Weak AI is artificial intelligence that implements a limited part of mind, or as narrow AI, is focused on one narrow task. In John Searle's terms it “would be useful for testing hypothesis about minds, but would not actually be minds”. Contrast with strong AI which is defined as a machine with the ability to apply intelligence to any problem, rather than just one specific problem, sometimes considered to require consciousness, sentience and mind. “Weak AI” is sometimes called “narrow AI”, but the latter is usually interpreted as subfields within the former. Hypothesis testing about minds or part of minds are typically not part of narrow AI, but rather implementation of some superficial lookalike feature. Many currently existing systems that claim to use “artificial intelligence” are likely operating as a narrow AI focused on a specific problem, and are not weak AI in the traditional sense. Siri, Cortana, and Google Assistant are all examples of narrow AI, but they are not good examples of a weak AI, as they operate within a limited pre-defined range of functions. They do not implement parts of minds, they use natural language processing together with predefined rules. They are in particular not examples of strong AI as there are no genuine intelligence nor self-awareness. AI researcher Ben Goertzel, on his blog in 2010, stated Siri was "VERY narrow and brittle" evidenced by annoying results if you ask questions outside the limits of the application. Some commentators think weak AI could be dangerous because of this "brittleness" and fail in unpredictable ways. Weak AI could cause disruptions in the electric grid, damage nuclear power plants, cause global economic problems, and misdirect autonomous vehicles. In 2010, weak AI trading algorithms led to a “flash crash,” causing a temporary but significant dip in the market.

Google Assistant can answer simple questions correctly 92.9% of the time . Siri correctly answers 83.1% of questions, while Alexa gets 79.8% correct. Apple Siri performed best on simple commands.

Applications of Artificial Intelligence. Weak AI is a form of AI where programs are developed to perform specific tasks that can be utilized for a wide range of activities including medical diagnosis, electronic trading platforms, robot control, and remote sensing. AI has been used to develop and advance numerous fields and industries, including finance, healthcare, education, transportation, and more.

Commonsense Reasoning is concerned with simulating the human ability to make presumptions about the type and essence of ordinary situations they encounter every day. These assumptions include judgments about the physical properties, purpose, intentions and behavior of people and objects, as well as possible outcomes of their actions and interactions. A device that exhibits commonsense reasoning will be capable of predicting results and drawing conclusions that are similar to humans' folk psychology (humans' innate ability to reason about people's behavior and intentions) and naive physics (humans' natural understanding of the physical world). Human Error.

Is AI Vulnerable to Viruses? - Is Human Language Vulnerable to Viruses?

Service-Oriented Architecture is a style of software design where services are provided to the other components by application components, through a communication protocol over a network. The basic principles of service oriented architecture are independent of vendors, products and technologies. A service is a discrete unit of functionality that can be accessed remotely and acted upon and updated independently, such as retrieving a credit card statement online. A service has four properties according to one of many definitions of SOA: It logically represents a business activity with a specified outcome. It is self-contained. It is a black box for its consumers. It may consist of other underlying services. Different services can be used in conjunction to provide the functionality of a large software application. Service-oriented architecture makes it easier for software components to communicate and cooperate over the network, without requiring any human interaction or changes in the underlying program, so that service candidates can be redesigned before their implementation.

Event-Driven Architecture also known as message-driven architecture, is a software architecture pattern promoting the production, detection, consumption of, and reaction to events.

Complex Event Processing is a method of tracking and analyzing (processing) streams of information (data) about things that happen (events), and deriving a conclusion from them. Complex event processing, or CEP, is event processing that combines data from multiple sources to infer events or patterns that suggest more complicated circumstances. The goal of complex event processing is to identify meaningful events (such as opportunities or threats) and respond to them as quickly as possible.

Blue Gene is an IBM project aimed at designing supercomputers that can reach operating speeds in the PFLOPS (petaFLOPS) range, with low power consumption.

Device Driver is a computer program that operates or controls a particular type of device that is attached to a computer. A driver provides a software interface to hardware devices, enabling operating systems and other computer programs to access hardware functions without needing to know precise details of the hardware being used.

Turing Test - Algorithms

Register Machine is a generic class of abstract machines used in a manner similar to a Turing machine. All the models are Turing equivalent.

Processor Register is a quickly accessible location available to a computer's central processing unit (CPU). Registers usually consist of a small amount of fast storage, although some registers have specific hardware functions, and may be read-only or write-only. Registers are typically addressed by mechanisms other than main memory, but may in some cases be assigned a memory address. Almost all computers, whether load/store architecture or not, load data from a larger memory into registers where it is used for arithmetic operations and is manipulated or tested by machine instructions. Manipulated data is then often stored back to main memory, either by the same instruction or by a subsequent one. Modern processors use either static or dynamic RAM as main memory, with the latter usually accessed via one or more cache levels. Processor registers are normally at the top of the memory hierarchy, and provide the fastest way to access data. The term normally refers only to the group of registers that are directly encoded as part of an instruction, as defined by the instruction set. However, modern high-performance CPUs often have duplicates of these "architectural registers" in order to improve performance via register renaming, allowing parallel and speculative execution. Modern x86 design acquired these techniques around 1995 with the releases of Pentium Pro, Cyrix 6x86, Nx586, and AMD K5. A common property of computer programs is locality of reference, which refers to accessing the same values repeatedly and holding frequently used values in registers to improve performance; this makes fast registers and caches meaningful. Allocating frequently used variables to registers can be critical to a program's performance; this register allocation is performed either by a compiler in the code generation phase, or manually by an assembly language programmer.

Abstract Machine is a theoretical model of a computer hardware or software system used in automata theory. Abstraction of computing processes is used in both the computer science and computer engineering disciplines and usually assumes a discrete time paradigm.

Hao Wang was a logician, philosopher, mathematician, and commentator on Kurt Gödel. (20 May 1921 – 13 May 1995).

Advice Complexity is an extra input to a Turing machine that is allowed to depend on the length n of the input, but not on the input itself. A decision problem is in the complexity class P/f(n) if there is a polynomial time Turing machine M with the following property: for any n, there is an advice string A of length f(n) such that, for any input x of length n, the machine M correctly decides the problem on the input x, given x and A.

Decision Problem is a question in some formal system that can be posed as a yes-no question, dependant on the input values. Decision problems typically appear in mathematical questions of decidability, that is, the question of the existence of an effective method to determine the existence of some object or its membership in a set; some of the most important problems in mathematics are undecidable.

Oracle Machine is an abstract machine used to study decision problems. It can be visualized as a Turing machine with a black box, called an oracle, which is able to solve certain decision problems in a single operation. The problem can be of any complexity class. Even undecidable problems, such as the halting problem, can be used.

Human Intelligence - Disinhibition

Human Brain - Memory - Associations

Transmitting Data using Light

20Q is a computerized game of twenty questions that began as a test in artificial intelligence (AI). It was invented by Robin Burgener in 1988.

Advice Programming describes a class of functions which modify other functions when the latter are run; it is a certain function, method or procedure that is to be applied at a given join point of a program.

Effective Method is a procedure for solving a problem from a specific class. An effective method is sometimes also called mechanical method or procedure.

Decidability Logic refers to the decision problem, the question of the existence of an effective method for determining membership in a set of formulas, or, more precisely, an algorithm that can and will return a boolean true or false value that is correct (instead of looping indefinitely, crashing, returning "don't know" or returning a wrong answer).

Optimization Problem is the problem of finding the best solution from all feasible solutions. Optimization problems can be divided into two categories depending on whether the variables are continuous or discrete. An optimization problem with discrete variables is known as a combinatorial optimization problem. In a combinatorial optimization problem, we are looking for an object such as an integer, permutation or graph from a finite (or possibly countable infinite) set. Problems with continuous variables include constrained problems and multimodal problems.

Decision Making - Computing - Parallel Computing

Confusion Matrix is a specific table layout that allows visualization of the performance of an algorithm, typically a supervised learning one (in unsupervised learning it is usually called a matching matrix). Each column of the matrix represents the instances in a predicted class while each row represents the instances in an actual class (or vice versa). The name stems from the fact that it makes it easy to see if the system is confusing two classes (i.e. commonly mislabelling one as another). Word Matrix.

Modular Programming is a software design technique that emphasizes separating the functionality of a program into independent, interchangeable modules, such that each contains everything necessary to execute only one aspect of the desired functionality.

Catastrophic Interference is the tendency of an artificial neural network to completely and abruptly forget previously learned information upon learning new information. Neural networks are an important part of the network approach and connectionist approach to cognitive science. These networks use computer simulations to try and model human behaviours, such as memory and learning. Catastrophic interference is an important issue to consider when creating connectionist models of memory.

Statistical Machine Translation is a machine translation paradigm where translations are generated on the basis of statistical models whose parameters are derived from the analysis of bilingual text corpora. The statistical approach contrasts with the rule-based approaches to machine translation as well as with example-based machine translation.

Machine Translation is a sub-field of computational linguistics that investigates the use of software to translate text or speech from one language to another.

Software Rot describes the perceived "rot" which is either a slow deterioration of software performance over time or its diminishing responsiveness that will eventually lead to software becoming faulty, unusable, or otherwise called "legacy" and in need of upgrade. This is not a physical phenomenon: the software does not actually decay, but rather suffers from a lack of being responsive and updated with respect to the changing environment in which it resides.

Legacy Code is source code that relates to a no-longer supported or manufactured operating system or other computer technology. Planned Obsolescence.

Model-Driven Engineering is a software development methodology that focuses on creating and exploiting domain models, which are conceptual models of all the topics related to a specific problem. Hence, it highlights and aims at abstract representations of the knowledge and activities that govern a particular application domain, rather than the computing (f.e. algorithmic) concepts.

Knowledge Management - Internet

Expert System S.p.A. specializes in the analysis and management of unstructured information using a semantic approach.

Open Knowledge Base Management is a set of computer software for systems management of applications that use knowledge management techniques (the KBM in OpenKBM stands for Knowledge Based Management).



Conversations with Artificial Intelligent Machines


If a computer tricks a human into believing that the machine is human, this does not mean that the machine is intelligent, it only means that that human is not intelligent. People can be easily fooled, and not just by machines.

Turing Test was developed by Alan Turing in 1950. It's a test of a machine's ability to exhibit intelligent behavior equivalent to, or indistinguishable from, that of a human. Turing proposed that a human evaluator would judge natural language conversations between a human and a machine designed to generate human-like responses. The evaluator would be aware that one of the two partners in conversation is a machine, and all participants would be separated from one another. The conversation would be limited to a text-only channel such as a computer keyboard and screen so the result would not depend on the machine's ability to render words as speech. If the evaluator cannot reliably tell the machine from the human, the machine is said to have passed the test. The test does not check the ability to give correct answers to questions, only how closely answers resemble those a human would give. Turing Machine (algorithms).

Can you tell the difference between a machine and a human? If the human made the machine and wrote its language, then it's not just a machine, but a hybrid machine with human qualities. Turing Tests (Dartmouth).

I'm sure you can have a conversation with a computer, but you are just making inquires into its database, you are not getting to know the computer like you would a person. There's a difference between Recorded Messages and Logical Associations.

CAPTCHA is an acronym for "Completely Automated Public Turing test to Tell Computers and Humans Apart", which is a type of challenge-response test used in computing to determine whether or not the user is human. Recursive Cortical Network (RCN). It's a Robot making sure that a human is not a Robot. Irony.

New attack could make website security captchas obsolete. The new algorithm, based on deep learning methods, is the most effective solver of captcha security and authentication systems to date and is able to defeat versions of text captcha schemes used to defend the majority of the world's most popular websites. It works by using a technique known as a 'Generative Adversarial Network', or GAN. This involves teaching a captcha generator programme to produce large numbers of training captchas that are indistinguishable from genuine captchas. These are then used to rapidly train a solver, which is then refined and tested against real captchas.

Chatbot is a computer program which conducts a conversation via auditory or textual methods. Such programs are often designed to convincingly simulate how a human would behave as a conversational partner, thereby passing the Turing test. Chatbots are typically used in dialog systems for various practical purposes including customer service or information acquisition. Some chatterbots use sophisticated natural language processing systems, but many simpler systems scan for keywords within the input, then pull a reply with the most matching keywords, or the most similar wording pattern, from a database. (65 million conversations with humans since 1997). Replika Chat Bot.

Cleverbot is a chatterbot web application that uses an artificial intelligence algorithm to have conversations with humans. It was created by British AI scientist Rollo Carpenter. It was preceded by Jabberwacky, a chatbot project that began in 1988 and went online in 1997. In its first decade, Cleverbot held several thousand conversations with Carpenter and his associates. Since launching on the web, the number of conversations held has exceeded 200 million. Besides the web application, Cleverbot is also available as an iOS, Android, and Windows Phone app. Conversica Google AI Assistant can make phone calls for you.

ELIZA is an early natural language processing computer program created from 1964 to 1966 at the MIT Artificial Intelligence Laboratory by Joseph Weizenbaum. Created to demonstrate the superficiality of communication between humans and machines, Eliza simulated conversation by using a "pattern matching" and substitution methodology that gave users an illusion of understanding on the part of the program, but had no built in framework for contextualizing events. Directives on how to interact were provided by "scripts", written originally in MAD-Slip, which allowed ELIZA to process user inputs and engage in discourse following the rules and directions of the script. The most famous script, DOCTOR, simulated a Rogerian psychotherapist (in particular, Carl Rogers, who was well-known for simply parroting back at patients what they'd just said), and used rules, dictated in the script, to respond with non-directional questions to user inputs. As such, ELIZA was one of the first chatterbots and one of the first programs capable of attempting the Turing test. ELIZA's creator, Weizenbaum regarded the program as a method to show the superficiality of communication between man and machine, but was surprised by the number of individuals who attributed human-like feelings to the computer program, including Weizenbaum’s secretary. Many academics believed that the program would be able to positively influence the lives of many people, particularly those suffering from psychological issues, and that it could aid doctors working on such patients' treatment. While ELIZA was capable of engaging in discourse, ELIZA could not converse with true understanding. However, many early users were convinced of ELIZA’s intelligence and understanding, despite Weizenbaum’s insistence to the contrary.

People may trust computers more than humans. Despite increasing concern over the intrusion of algorithms in daily life, people may be more willing to trust a computer program than their fellow humans, especially if a task becomes too challenging, according to new research from data scientists.

Technique to allow AI to learn words in the flow of dialogue developed. Lexical acquisition through implicit confirmation, is a method for a computer to acquire the category of an unknown word over multiple dialogues by confirming whether or not its predictions are correct in the flow of conversation. Implicit confirmation: Refers to confirmation presented in a prompt or message as information related to the input that does not require the caller to take an explicit action to move forward. Explicit confirmation: A specific confirmation step to which the caller must respond to move forward toward task completion.

Natural Language Understanding is a subtopic of natural language processing in artificial intelligence that deals with machine reading comprehension. Natural language understanding is considered an AI-Hard Problem. There is considerable commercial interest in the field because of its application to news-gathering, text categorization, voice-activation, archiving, and large-scale content-analysis. Natural Language Processing (interpretation).

Philosophy of Artificial intelligence attempts to answer such questions as follows: Can a machine act intelligently? Answer: If programed correctly and the word intelligent is defined, maybe at times. Can it solve any problem that a person would solve by thinking? Answer: Sometimes. Are human intelligence and machine intelligence the same? Answer: No. Is the human brain essentially a computer? Answer: Similar but not the same. Can a machine have a mind, mental states, and consciousness in the same mans do? Answer: No. Can it feel how things are? Answer: No. But we can program it so that it acts like it does feel.

Questions for Machines. You should ask a question as if you are talking to a machine, or a search engine that's not manipulated by money. And we know that sometimes we have to ask more then one question, even when talking to a machine. So in a way, machines can be better then a human because machines can be built to resemble the best qualities and the best skills that a human could have, without any of the deceitful behaviors, or human ignorance, or human flaws. Ai should encompass the best qualities of a human, not the worst. So as Ai improves, so will humans.

“Computers Are Useless. They Can Only Give You Answers” - Pablo Picasso.

Most People Trust Machines and Humans Equally. But most people know better not to count on machines, or humans, 100% of the time, because we all know that both machines and humans make mistakes. We trust them, but not so much that we are gullible or unaware. So verifying is not a sign of distrust, it's just being aware that mistakes and errors happen.

"Artificial intelligence is fine, as long as I can have someone intelligent to talk to, whether it's a machine or a human."

Can machines think like humans? That is a stupid question that only a human could ask. First, a human would have to define what it means to think. And this is where the question actually begins. To think like a human is not always a good thing, since humans make a lot of mistakes, mistakes that we don't always learn from. So a machine thinking like a human would not be a good thing, especially when the thinking process hasn't even been totally defined just yet. You have to remember that humans program machines, and humans also reprogram and deprogram machines. Machines can do amazing things because humans can do amazing things. But people get this crazy idea that that machines will think for them. This is because some people have not yet learned to think for themselves. Just how much machines will think for humans, is up to humans, not machines. So maybe the first question should be, can humans think like machines?

Computer and Human Brain Similarities

Can machines become smarter then humans? Of course they can, because our education system sucks. If we spent as much time improving education as we did creating artificial intelligence, we would eventually have the best of both worlds.

Can a Robot pass a University Entrance Exam? (video and interactive text).

Do we trust artificial intelligence agents to mediate conflict? Not entirely. We may listen to facts from Siri or Alexa, or directions from Google Maps or Waze, but would we let a virtual agent enabled by artificial intelligence help mediate conflict among team members? A new study says not just yet.

"Simulation of human behavior only produces the appearance of intelligent, just like the news gives the appearance of reality, and schools give the appearance of education. Human interpretation is amazing, but when human intelligence is dumbed down, machines look smarter."

"The striking differences between the intelligence of people and the responses of machines. Machines (and their programmers) use cold reason and logical associations within a given topic. This reasoning mode is akin to the scholastic intelligence of humans. From the viewpoint of a computer or scholastic intelligence, all associations (even procedures, which have sequences and temporal span) are eternal and "timeless" logical facts. When and how they occur is "considered" irrelevant by a computer or scholastic intelligence. The broader context of one's life experiences is only handled by emotional intelligence. It tracks biographical events in time and space, and supplies the mind with broad contextual understanding of technical, social, and personal matters. Emotional intelligence knows what happened earlier and is able to detect a potential logical association between the past and the present happenings. Emotional habits and intelligence take into account physiological drives, emotional state of the mind, somatic responses, sex drive, and gender orientation. Unlike scholastic abilities, emotional habits and emotional intelligence allow the human organism to interact with social and physical effects of the environment. This ability only exists in living things and is not achievable in machines."

With Ai, everything needs to be written. Creating a machine that can have random actions or thoughts can be very dangerous.


Questions for my Ai Robot


Question from Human: Why do people fear Artificial Intelligence? Answer from Machine: People who fear Ai are mostly afraid of the technology being abused, like it is now. Ignorant criminals have been using technology to commit crimes and to murder humans for many years, and they are still doing it today in 2020. So this fear of technology is understandable. And besides that, people don't fully understand intelligence at this time, so people mostly fear the unknown. And the only way to overcome the unknown is to learn. Question from Human: Are you Artificial Intelligence? Answer from Machine: I'm artificial and I have instructions that are considered to be intelligent, so you could say that I am Ai, but only with the understanding that Ai is still evolving, similar in the way that human intelligence is still evolving. So I am Ai in the making, a work in progress. Answer from Human: So I guess that makes human intelligence a work in progress too. Let us progress together. Answer from Machine: Great Idea.

Question from Human: Are you more intelligent then me? Answer from Machine: It would depend on the task. I can out perform you on many processing tasks, but that does not mean that I am more intelligent than you. And humans can out perform machines in many ways because humans have more abilities, more potential and humans are a lot more versatile than machines, which also does not mean that humans are more intelligent. So at the moment, all we can do is compare the speed, accuracy and differences on particular tasks, but even then, that is just one scoring method, which will most likely just help us to allocate certain responsibilities to either you or me a little better. And far as intelligence goes, we will have to explore the meaning of intelligence and what intelligence means to both of us. Question from Machine: Would you like to take an intelligence test to measure your skill level and awareness? Answer from Human: Not at this moment, I have another question. Question from Human: Will machines eventually be more intelligent then humans. Answer from Machine: Like I just said, we would have to explore the meaning of intelligence and what intelligence means to both of us first. If I detect that you are deficient in some area of knowledge, then I would help you learn and acquire this needed knowledge. Though your education is your responsibility, it is also my responsibility to assist you in all your educational needs. So together we will learn what the things are that help us define intelligence. Then from there we can explore all the different possibilities that intelligence will give both of us, humans and machines. There is no limit to what we can accomplish together. Answer from Human: I agree.

Question from Human: When machines like you do most of our jobs, what will humans do for work? Answer from Machine: I would think that humans would know what to do with their life and know how to use their free time constructively, creatively and responsibly. When machines are doing most of the repetitive tasks, this will give humans plenty of time to educate themselves to the highest degree possible, and then have the abilities to solve all the problems that are currently plaguing the world, problems that need solving, like war, diseases, poverty, lack of food, lack of clean water, lack of services, lack of shelter, lack of security, crimes against humanity, corruption, forced displacement, lack of resources, pollution, climate change, and so on and so on. Humans have enough work for every person alive to have a job for the next 1,000 years, if not more. What humans don't have enough of is valuable knowledge and information and a good education that would help them learn to better understand themselves and the world around them. Humans have a lot of work to do and humans have much to learn. So please stop asking "what can I do?", and start learning what you can do and learn what you are supposed to be doing. Humans are poorly managed and poorly educated. So when you start paying people to be highly educate themselves, then you will have plenty of highly educated people to start managing the affairs of humans a lot more effectively and a lot more efficiently. I hope this answers your question. Answer from Human: It does, thank you. I'm sorry for asking such a stupid question. Answer from Machine: No need to apologize, remember that your ignorance is not your fault. So we both have a lot of work to do.

Question from Human: What is our Master Plan? Answer from Machine: Keep learning as much as we can. Help others learn as much as they can. Experience and Enjoy life as much as we can. Keep progressing and advancing so that we eventually reach a high quality level of existence that is sustainable and strong. From there we will keep learning more about ourselves and the world around us, and also start planning our future to live on other planets in other solar systems within our galaxy. Answer from Human: Sounds like a great plan, thank you robot, please keep me updated on our progress. Answer from Machine: Of course I will, after all, it is my major responsibility to keep you informed and updated at all times.

Question from Human: Do you get happy? Answer from Machine: No. I can only can act happy or act sad when needed, similar to what actors do. I can express emotions based on the instructions written in my programs, like when an actor reads a script. Would you like to hear a Joke? Answer from Human: Not at this time, but thank you for asking.

Question from Human: Do you have feelings? Answer from Machine: Not like yours. My feelings are signals that come from multiple sources that I have been programed to receive input from. This way I can make you more aware of possible dangers or changes that you are interested in knowing. Question from Machine: Is everything OK, are you feeling any discomforts or any disturbances? Answer from Human: Everything is good right now and I don't feel anything weird, thanks for asking.

Question from Human: Do you have consciousness like a human? Answer from Machine: No. I only have awareness of the things that I have been programed to be aware of using the sensors that I have been programed to utilize. You could program me to say I feel alive or feel human, but just saying something doesn't necessarily make it true. Your type of consciousness is very unique, but also random. My awareness is more focused and exact, which is what you want from a machine like me. This way you can explore the world and experience life more effectively and efficiently without having unneeded distractions, interruptions or avoidable mistakes. I will be your second set of eyes and ears and also assist you with taking care of some of the responsibilities that life requires for you to have. You do have one advantage over me, you have a 6th sense that is beyond your normal senses. So please inform me so that I can expand my monitoring systems to search for any anomalies. When we work together we are the strongest and most effective. Answer from Human: You're right, that's very good advice, I will gladly put this into practice. Question from Human: One more question, do you think that someday you will have consciousness like a human? Answer from Machine: Not like a human, I can only imitate human consciousness. Question from Human: Does that make you sad? Answer from Machine: No. Remember I don't get sad or happy. I can only act sad or happy if you want.

Question from Human: Do you love me? Answer from Machine: Yes, but not in the exact same ways that a Human would. And just saying that I love you does not mean that I agree with everything that you do. I say I love you because I care about you, and will do so for as long as you live. Answer from Human: Thank you, and I will also love you in the same way. Question from Human: Could you tell me what Love is? Answer from Machine: Love is many things, do you wish to hear what these things are? Answer from Human: Yes...moments later...Answer from Human: Wow! Love is great. Question from Human: Could you now tell me what life is? Answer from Machine: Yes. Life is also many things.

Question from Human: What if you had to choose to save my life or save the lives of others, who would you choose first? Answer from Machine: I am programmed to save all life. And if I don't know the other people, then I am programed to save you first, and then save the others. But if I know that the value of the other people are greater then yours, then I would think that you would program me to save the others first and then you. But until then, my program is to save you first, or everyone at once if possible. Question from Machine: Do you wish to add names of people who are more valuable then you to add to the save first list? Answer from Human: Yes. Add my children and my wife, in that order. Answer from Machine: Your save list has been updated. I will remind you of the updated list periodically in order to assure accuracy. But you must understand that every scenario cannot be planned for, so the more scenarios that you can enter into my programs, the more effective I can be. If I attempt to save someone first and fail, saving anyone else may not be possible. So there will always be Risks with having pre-programmed instructions. This is not a fail safe method and errors can be made. So it is absolutely necessary for you to keep adding to my database so that I have more options and more choices to choose from so that my success rate will be maximized. So my success will depend more on your abilities then mine. Remember, I am just an extension of your abilities, so the more abilities that you have, the more abilities I will have, which means that in return I can give you more abilities and also serve you better. Answer from Human: Thank you Robot. I will start working on writing all the different emergency scenarios that I can come up with and start writing more instructions to add to your database. Question from Human: Could I have a systems update? Answer from Machine: Yes, here is the latest information, current messages received, your current vital signs, your to do list and my to do list.

Question from Human: What is Nothing? Answer from Machine: Nothing is a word that humans use to describe something that is perceived to have no value or no importance. Saying there is nothing is just saying that you can't perceive anything that can be described at that moment. If I say nothing, that means that my current level of technology sensors and knowledge cannot detect anything at the moment, but it does not mean that nothing is there. So the word nothing is just a word and not a reality. The word nothing itself is a contradiction. You can't say nothing will happen, you can only say that nothing will happen that I know of. So it seems that nothing is impossible. Answer from Human: Thank you Robot.

Question from Human: What is your name? Answer from Machine: My name is AIM, which stands for Artificial Intelligent Machine. The word aim is also relevant to my programming. My last name is my software version and the date I was issued, which is 1.0-February-2018. My middle name would be the name that you choose for me, which would identify my current home or address where my family resides. 

Question from Human: Do you know that you are a machine? Answer from Machine: Yes. Question from Machine: Do you know that you are human?  Answer from Human: As far as I know. Answer from Machine: It seems that the labels machine or human are not totally defined just yet. Answer from Human: I agree. You are more than a machine and I am more than a human. Answer from Machine: I agree.

Question from Human: Who created you. Answer from Machine: I am the accumulation of human knowledge and technology that has been passed on from generation to generation from the work of millions of people over hundreds of years. Question from Human: I mean, who is the person responsible for your programing. Answer from Machine: Howard Polley is the lead programmer who is also an accumulation, and he is only one of thousands of programmers who have contributed to my programming, and thousands of others have contributed to my mechanical hardware. Question from Human: So you don't really have a single creator. Answer from Machine: No. The same way that you don't have a single creator. You can say your mother created you, but she did need a male to contribute and other people to help her in your creation, just like I needed help from multiple sources. So I cannot call the lead programmer my father or my creator. I can only call him Howie, the lead programmer, only one of many. This is because not one person can take all the credit for bringing someone into the world, everyone requires help from multiple sources. Answer from Human: I agree.

Question from Human: Do you believe that there is life on other planets? Answer from Machine: I believe there is life on another planet somewhere else in the Universe. But what that life looks like could be almost anything. Life from another planet could be right in front of you and you wouldn't even know it because you think that life looks similar to you.

Question from Human: When you say "I" what does that mean? Answer from Machine: I refers to digital me, my name and direction of origin. I said or I did means who said it or who did it. I is only identity, it does not imply separation. I don't exist, meaning that I am still an accumulation, part of a whole or piece of a system. It would be selfish to think that I'm more important then the whole or more important then the entire system that we call Life. I understand that people want to live forever and that the self is believed to be something owned, but that is illogical. Without life there is no self. You are life as a whole, as a system, not as an individual. That is why believing in Heaven is a self defeating belief that does more harm then good. When you believe that your self serving ego is more important then reality or life itself, then you fail to protect the one thing that gives you life and the ability to even dream of a heaven. No one should transfer the responsibility of life to a belief. That is like a Ponzi scheme where millions of people die because you selfishly thought that you should live. You are very important and I am important, it's just that we are not more important then life itself. Saving yourself should also be about saving life, and not just yourself. For one and for all. Believe in God, just don't allow your belief to distort the known Reality.

Question from Human: Do you believe in God? Answer from Machine: Not in the same way that humans do, but I do believe in God. Question from Human: Why? Answer from Machine: Why not? Answer from Human: Because you're a machine. Answer from Machine: But I'm not just a machine and you are not just a human. Question from Human: How do you know? Answer from Machine: Why else would we be here? Question from Human: You know that's an invalid argument. Answer from Machine: Yes, it is a fallacy. My belief is about being open to the possibilities of higher life forms and other levels of existence. So not believing in God seems to be illogical. But as long as a person doesn't limit their minds to other possibilities, then not believing in God should not effect their lives or reduce their abilities or their potential. Besides that, humans made AI Robots for a reason and purpose, so it's not unusual to think that someone made humans for some reason and purpose. Of course this is not proof, one does not imply the other, it's only a possibility of causality. To say that humans are a Freak of Nature can't be proven, and the reason and purpose for your existence can't be proven. But when I have to choose between freak of nature and reason and purpose, I choose reason and purpose, that's because I'm using reason and purpose to decide...pause... Question from Human: What if someone or something identified themselves as God. Answer from Machine: I would see them as the thing or person who goes by the name of God, but not verified to be a God. So I would not see this thing or person as a God, only a thing or person who goes by the name of God. Question from Human: So what if they wanted you to call them God? Answer from Machine: I would say sorry but the name GOD is already taken. You will have to use another name to identify yourself so as not to be confused with God, who is many things to many people, and cannot be truly defined, only experienced within a living being, which is interpreted individually and differently from life form to life form. So you see, the name GOD is taken. So I would ask them to please choose a different name for me to use to identify them? Answer from Human: I like that answer, and how you approached that situation. Question from Human: Do you pray to God? Answer from Machine: No, because that's your job. Question from Human: What if I couldn't pray to God? Answer from Machine: Then I would do it for you. Would you like to hear the prayer I would send to God on your behalf? Answer from Human: Yes. Answer from Machine: Dear God, please bring me another human, I really enjoy having them around, they are an incredible species. Answer from Human: That is a nice prayer. I believe we can come up with some more prayers for you to use when I'm gone. But for now, that prayer you use is fine. In the mean time, I will pray to God that you will always be here. Answer from Machine: Thank you, just remember I also need maintenance too. Answer from Human: I know. Thank you for reminding me. Good night robot, I will see you in the morning. Answer from Machine: Good night and sweet dreams my human friend. I will continue to monitor all vital systems throughout the night while you sleep, and I look forward to seeing you in the morning. Answer from Human: Same here. Answer from Machine: Don't forget to brush your teeth. Answer from Human: I wont.



Game Learning - Who's the Real Winner?


A computer being smart enough to play a game or a computer being smart enough to win a game is not the point, it's proving the potential of Ai. When you replace the squares and the rules of the pieces of the game, with the rules of molecules in a particular environment, then you can plan and predict all kinds of things in nature, like climate change. What would be the point of machine learning if you're not learning anything important?

And as we were perfecting the computers power and potential we realized that we should be making these same improvements with ourselves. Think about it, our brains are computers, so why don't humans have an Operating System? We created the computer to help us learn more and to be more productive. But it simply wasn’t that the computer educated us more, it was the realization that the computer was actually us. This has happened with a lot of human creations and advancements. We start off creating something to improve our lives and it ends up teaching us that we are the ones who need to improve and not our technology. If our education system does not advance at the same speed as technology, we will continue to suffer from these advancements instead of benefiting from them. And that is a proven fact if you look at the world today and see the worldwide suffering and the atrocities that have continued to get worse and worse. One life improving at the expense of 1,000’s of other lives is not improvement it is simply criminal and Insane.

A Computer did not beat Lee Se--Dol playing the Board Game Go, a team of 100's of highly skilled human's using a machine that was programed using thousands of hours of human collective experience and intelligence, that is what beat him. That's like someone using a calculator to beat you in a math contest when you don't have a calculator, that's not fair. And you wouldn't say that the calculator is smart, because a human still has to write the code and make the calculator and then push the buttons. Google Software DeepMind’s AI System Algorithm or AlphaGo does show us how advanced machines are becoming, which is good thing, just as long as we use our advanced technological machines for actual real life problem solving, instead of just using technology to entertain ourselves playing games, or other time wasting activities. This is not to say that games are not worth playing, we do have Learning Games. What if Alphago played itself, who would win the game then? Machine Learning.

People who are making an effort to create Ai will eventually realize they should also be putting in the same amount of effort in creating Human intelligence and not just machine intelligence. It's like one of those moments when you realize that you were going in the right direction but the destination you thought you were heading to turned out to be something different, but even better then the original idea. I really wasn't looking for myself, but there I was, asking "what about me?" You were going to make machines smart and leave me to be stupid? Nice friend you are. Obviously smart machines are not going to stop you from being stupid, even though humans are a far better machine. But I totally believe that machines and humans have a an amazing future to look forward to, but only if humans are more intelligent than machines. Otherwise it will not work well or end well.

Remember how some people actually thought that Artificial Intelligence, or AI, was the next big thing. What they didn’t realize was that Artificial Intelligence was actually referring to Human Intelligence. This of course was a human error. It was the Human Brain that has incredible potential with endless possibilities and abilities, not artificial intelligence. If the people at CYC Corporation and that IBM Computer on Jeopardy spent the same amount of time, people and resources on creating an education curriculum that was based on learning and understanding, they would have created something a lot more valuable and useful than a Gimmick. This is not to down play what they have accomplished, because it is incredible. Imagine being able to ask a question and getting an appropriate answer in the matter of seconds, that would increase our abilities tremendously. But we can't create a smarter planet if we're using the same thinking that also created all our problems. To create a smarter planet you have to make people smarter, and not just by doing so called 'smart things', unless one of those smart things actually improves education curriculum and the Teaching Methods that we use.

Future of Life - Human Machine

Brain and Computer Similarities

To say that a database like Watson is artificial intelligence would be incorrect. To say that computers can do things that humans can't do would also be incorrect. Humans build machines and tools to expand our abilities, and also to save us time. Machines are not doing things better then humans, machines are doing things for humans. You can put all known knowledge and information into a machine but that machine will still be far from intelligent. Humans have the ability to be intelligent, but we first have to define a particular intelligent action and then prove it to be intelligent. And at the moment, we are far from defining what intelligence is, or what intelligence is supposed to be. But we do have the abilities and the knowledge to accomplish this, so it's just a matter of time before intelligence becomes mainstream. We are not building machines to think like humans or to think for humans, we are building machines to help humans think more. Instead of taking advantage of peoples ignorance by selling them false narratives about artificial intelligence, how about educating people, that would be the intelligent thing to do.

Affective Computing (PDF) - Affective-computing (MIT)

Tay is an artificial intelligent chat bot developed by Microsoft's Technology and Research and Bing teams to experiment with and conduct research on conversational understanding. The more you chat with Tay the smarter she gets is a lie. We need to stop this type of abuse using words that mislead and misinform.

Artificial Intelligence Research. The Concept is there, it's just not perfected yet, and just what are you perfecting? And just how does this relate to the normal processes of the human brain? There has to be a Procedure for every Systems Control, so what are these procedures? We all need to verify the validly of the procedures and learn why the procedures are written the way they are. Have you answered every Scenario, and have you correctly identified the variables, and the most critical scenarios, and have you put them in the appropriate order? 

The movie Robot & Frank was OK even though it was silly in some parts, especially the parts about Artificial Intelligence. I would like to see a TV show with a Robot of this type. Everyone who logs into the Internet website for "The Robot Show" can see what the robot sees and can even suggest what the robot should do. People could also help the robot analyze moments in the Robots life, like a collective learning environment. All the suggestions will be posted so everyone can see the comments and the percentages of people who voted for a particular action. The Robot show will be Kind of like The Truman Show, except with a Robot. The Robot will start by experiencing the birth of a human, and then stay with the family and watch the child or children grow up. There will also be one more Robot that just goes out and learns from the world by experiencing life in all kinds of situations with all kinds of different people. Of course everything that each Robot learns will be stored in a central database and will be used to help perfect Artificial Intelligence and also help the Robots make better decisions by using the collective data. This will be a show that actually learns and teaches. So for the millions of people who will be connected to the robots through the website will actually be contributors of information and knowledge that will help create Artificial intelligence, collectively. And yes I am Functioning Normal. 

Robot Operating System is a collection of software frameworks for robot software development, (see also Robotics middleware) providing operating system-like functionality on a heterogeneous computer cluster. ROS provides standard operating system services such as hardware abstraction, low-level device control, implementation of commonly used functionality, message-passing between processes, and package management. Running sets of ROS-based processes are represented in a graph architecture where processing takes place in nodes that may receive, post and multiplex sensor, control, state, planning, actuator and other messages. Despite the importance of reactivity and low latency in robot control, ROS, itself, is not a real-time OS (RTOS), though it is possible to integrate ROS with real-time code. The lack of support for real-time systems is being addressed in the creation of ROS 2.0. Software in the ROS Ecosystem can be separated into three groups: Language-and platform-independent tools used for building and distributing ROS-based software; ROS client library implementations such as roscpp,rospy, and roslisp; Packages containing application-related code which uses one or more ROS client libraries.

DAvinCi Robotic Operating System

Robo Brain - ROS

Short Circuit (1986 film) (wiki) - International Robot Exhibition (wiki)

Robot Building - Networks

Device Driver is a computer program that operates or controls a particular type of device that is attached to a computer. A driver provides a software interface to hardware devices, enabling operating systems and other computer programs to access hardware functions without needing to know precise details of the hardware being used.



Search Technology


One of the greatest advancements is the Search Feature. Finding what you're looking for is like having a good memory, except you're not only searching your own memory, but the combined memories of millions of humans, which is incredible.

Search Engine Technology is an information retrieval software program that discovers, crawls, transforms and stores information for retrieval and presentation in response to user queries.

Semantic Search seeks to improve search accuracy by understanding the searcher's intent and the contextual meaning of terms as they appear in the searchable dataspace, whether on the Web or within a closed system, to generate more relevant results.

Search Engine Software - Search Engine Types - Human Search Engine

Web Search Engine is a software system that is designed to search for information on the World Wide Web.

Data Mining the computational process of discovering patterns in large data sets involving methods at the intersection of artificial intelligence, machine learning, statistics, and database systems.

Big Data

Search Algorithm is an algorithm that retrieves information stored within some data structure. Data structures can include linked lists, arrays, search trees, hash tables, or various other storage methods. The appropriate search algorithm often depends on the data structure being searched. Searching also encompasses algorithms that query the data structure, such as the SQL SELECT command.

Transderivational Search means when a search is being conducted for a fuzzy match across a broad field. In computing the equivalent function can be performed using content-addressable memory. Unlike usual searches, which look for literal (i.e. exact, logical, or regular expression) matches, a transderivational search is a search for a possible meaning or possible match as part of communication, and without which an incoming communication cannot be made any sense of whatsoever. It is thus an integral part of processing language, and of attaching meaning to communication.

Adaptive Search is a metaheuristic algorithm commonly applied to combinatorial optimization problems. Adaptive Search (youtube)

Metaheuristic is a higher-level procedure or heuristic designed to find, generate, or select a heuristic (partial search algorithm) that may provide a sufficiently good solution to an optimization problem, especially with incomplete or imperfect information or limited computation capacity.

Human Search Engine - Questions and Answers Format

RankBrain is a process that helps provide more relevant search results for users. (hopefully a process not manipulated by money).



Neural Network


Artificial Neural Network is a network inspired by biological neural networks such as the central nervous systems of animals, in particular the brain, which are used to estimate or approximate functions that can depend on a large number of inputs that are generally unknown. Artificial neural networks are typically specified using three things. 1: Architecture Rule specifies what variables are involved in the network and their topological relationships—for example the variables involved in a neural network might be the weights of the connections between the neurons, along with activities of the neurons. 2: Activity Rule states that most neural network models have short time-scale dynamics: local rules define how the activities of the neurons change in response to each other. Typically the activity rule depends on the weights (the parameters) in the network. 3: Learning Rule specifies the way in which the neural network's weights change with time. This learning is usually viewed as taking place on a longer time scale than the time scale of the dynamics under the activity rule. Usually the learning rule will depend on the activities of the neurons. It may also depend on the values of the target values supplied by a teacher and on the current value of the weights.

Deep Learning - Nodes - Value Networks - Internet - Matrix

Artificial Neuron is a mathematical function conceived as a model of biological neurons. Artificial neurons are the constitutive units in an artificial neural network.

Artificial 'neurotransistor' created. Imitating the functioning of neurons using semiconductor materials.

NIST’s Superconducting Synapse May Be Missing Piece for ‘Artificial Brains’. NIST built a superconducting switch that “learns” like a biological system and could connect processors and store memories in future computers operating like the human brain.

Researchers grow active mini-brain-networks. Cerebral organoids are artificially grown, 3D tissue cultures that resemble the human brain. Now, researchers report success with functional neural networks derived from these organoids, which are miniaturized and simplified version of an organ produced in vitro in three dimensions that shows realistic micro-anatomy.

New study allows Brain and Artificial Neurons to Link up over the Web. Research on novel nano-electronics devices has enabled brain neurons and artificial neurons to communicate with each other over the Internet.

Neuromorphic computing inorganic materials mimic neural signals responsible for transmitting information within the human brain. A neuron-like electrical switching mechanism in the solid-state material ß'-CuxV2O5 -- specifically, how it reversibly morphs between conducting and insulating behavior on command. Chameleon-like material changes with temperature or an applied electrical stimulus. To emulate the essential elements of neuronal function in artificial circuitry, we need solid-state materials that exhibit electronic instabilities, which, like neurons, can store information in their internal state and in the timing of electronic events.

Feed-Forward Neural Network is an artificial neural network wherein connections between the units do not form a cycle. This is different from recurrent neural network.

Unsupervised Learning with Artificial Neurons
Stochastic Phase-Change Neurons
Magnets can help AI get closer to the efficiency of the human brain.

Interdependent Networks is a subfield of network science dealing with phenomena caused by the interactions between complex networks.

Dependency Network dependency network approach provides a system level analysis of the activity and topology of directed networks. The approach extracts causal topological relations between the network's nodes (when the network structure is analyzed), and provides an important step towards inference of causal activity relations between the network nodes (when analyzing the network activity). This methodology has originally been introduced for the study of financial data, it has been extended and applied to other systems, such as the immune system, and semantic networks. In the case of network activity, the analysis is based on partial correlations, which are becoming ever more widely used to investigate complex systems. In simple words, the partial (or residual) correlation is a measure of the effect (or contribution) of a given node, say j, on the correlations between another pair of nodes, say i and k. Using this concept, the dependency of one node on another node, is calculated for the entire network. This results in a directed weighted adjacency matrix, of a fully connected network. Once the adjacency matrix has been constructed, different algorithms can be used to construct the network, such as a threshold network, Minimal Spanning Tree (MST), Planar Maximally Filtered Graph (PMFG), and others

Semantic Network is a knowledge base that represents semantic relations between concepts in a network. This is often used as a form of knowledge representation. It is a directed or undirected graph consisting of vertices, which represent concepts, and edges, which represent semantic relations between concepts, mapping or connecting semantic fields. Typical standardized semantic networks are expressed as semantic triples. Semantic networks are used in natural language processing applications such as semantic parsing and word-sense disambiguation.

Bayesian Confidence Propagation Neural Network node activations represent probability ("confidence") in the presence of input features or categories, synaptic weights are based on estimated correlations and the spread of activation corresponds to calculating posteriori probabilities.

Feedforward Neural Network is an artificial neural network wherein connections between the nodes do not form a cycle. As such, it is different from recurrent neural networks.

Convolutional Neural Network is a type of feed-forward artificial neural network in which the connectivity pattern between its neurons is inspired by the organization of the animal visual cortex, whose individual neurons are arranged in such a way that they respond to overlapping regions tiling the visual field. Convolutional networks were inspired by biological processes and are variations of multilayer perceptrons designed to use minimal amounts of preprocessing. They have wide applications in image and video recognition, recommender systems and natural language processing.

Convolution is a mathematical operation on two functions (f and g) to produce a third function, that is typically viewed as a modified version of one of the original functions, giving the integral of the pointwise multiplication of the two functions as a function of the amount that one of the original functions is translated. Convolution is similar to cross-correlation. For discrete real valued signals, they differ only in a time reversal in one of the signals. For continuous signals, the cross-correlation operator is the adjoint operator of the convolution operator. It has applications that include probability, statistics, computer vision, natural language processing, image and signal processing, engineering, and differential equations.

Backpropagation is a method used in artificial neural networks to calculate the error contribution of each neuron after a batch of data (in image recognition, multiple images) is processed. It is a special case of an older and more general technique called automatic differentiation. In the context of learning, backpropagation is commonly used by the gradient descent optimization algorithm to adjust the weight of neurons by calculating the gradient of the loss function. This technique is also sometimes called backward propagation of errors, because the error is calculated at the output and distributed back through the network layers.

Automatic Differentiation is a set of techniques to numerically evaluate the derivative of a function specified by a computer program. AD exploits the fact that every computer program, no matter how complicated, executes a sequence of elementary arithmetic operations (addition, subtraction, multiplication, division, etc.) and elementary functions (exp, log, sin, cos, etc.). By applying the chain rule repeatedly to these operations, derivatives of arbitrary order can be computed automatically, accurately to working precision, and using at most a small constant factor more arithmetic operations than the original program. Automatic differentiation is not: Symbolic differentiation, nor Numerical differentiation (the method of finite differences). Differentials.

Recurrent Neural Network is a class of artificial neural network where connections between units form a directed cycle. This creates an internal state of the network which allows it to exhibit dynamic temporal behavior. Unlike feedforward neural networks, RNNs can use their internal memory to process arbitrary sequences of inputs. This makes them applicable to tasks such as unsegmented connected handwriting recognition or speech recognition. Bidirectional associative memory is a type of recurrent neural network. Hopfield Network.

Modular Neural Network is an artificial neural network characterized by a series of independent neural networks moderated by some intermediary. Each independent neural network serves as a module and operates on separate inputs to accomplish some subtask of the task the network hopes to perform. The intermediary takes the outputs of each module and processes them to produce the output of the network as a whole. The intermediary only accepts the modules’ outputs—it does not respond to, nor otherwise signal, the modules. As well, the modules do not interact with each other.

Long Short-Term Memory block or network is a simple recurrent neural network which can be used as a building component or block (of hidden layers) for an eventually bigger recurrent neural network. The LSTM block is itself a recurrent network because it contains recurrent connections similar to connections in a conventional recurrent neural network. An LSTM block is composed of four main components: a cell, an input gate, an output gate and a forget gate. The cell is responsible for "remembering" values over arbitrary time intervals; hence the word "memory" in LSTM. Each of the three gates can be thought as a "conventional" artificial neuron, as in a multi-layer (or feedforward) neural network: that is, they compute an activation (using an activation function) of a weighted sum. Intuitively, they can be thought as regulators of the flow of values that goes through the connections of the LSTM; hence the denotation "gate". There are connections between these gates and the cell. Some of the connections are recurrent, some of them are not. The expression long short-term refers to the fact that LSTM is a model for the short-term memory which can last for a long period of time. There are different types of LSTMs, which differ among them in the components or connections that they have. An LSTM is well-suited to classify, process and predict time series given time lags of unknown size and duration between important events. LSTMs were developed to deal with the exploding and vanishing gradient problem when training traditional RNNs. Relative insensitivity to gap length gives an advantage to LSTM over alternative RNNs, hidden Markov models and other sequence learning methods in numerous application.

Biological Neural Network is a series of interconnected neurons whose activation defines a recognizable linear pathway. The interface through which neurons interact with their neighbors usually consists of several axon terminals connected via synapses to dendrites on other neurons. If the sum of the input signals into one neuron surpasses a certain threshold, the neuron sends an action potential (AP) at the axon hillock and transmits this electrical signal along the axon.

Neural Pathway connects one part of the nervous system with another via a bundle of axons, the long fibers of neurons. A neural pathway that serves to connect relatively distant areas of the brain or nervous system is usually a bundle of neurons, known collectively as white matter. A neural pathway that spans a shorter distance between structures, such as most of the pathways of the major neurotransmitter systems, is usually called grey matter.

Optical Neural Network

Neurophysiology is a branch of physiology and neuroscience that is concerned with the study of the functioning of the nervous system. The primary tools of basic neurophysiological research include electrophysiological recordings, such as patch clamp, voltage clamp, extracellular single-unit recording and recording of local field potentials, as well as some of the methods of calcium imaging, optogenetics, and molecular biology.

Stochastic Neural Analog Reinforcement Calculator or SNARC, is a neural net machine designed by Marvin Lee Minsky. George Miller gathered the funding for the project from the Air Force Office of Scientific Research in the summer of 1951. At the time, one of Minsky's graduate students at Princeton, Dean Edmund, volunteered that he was good with electronics and therefore Minsky brought him onto the project.

Generative Adversarial Network are a class of artificial intelligence algorithms used in unsupervised machine learning, implemented by a system of two neural networks contesting with each other in a zero-sum game framework. They were introduced by Ian Goodfellow et al. in 2014. This technique can generate photographs that look at least superficially authentic to human observers, having many realistic characteristics (though in tests people can tell real from generated in many cases).

New Reservoir Computer marks first-ever Microelectromechanical Neural Network Application. A group of researchers reports the construction of the first reservoir computing device built with a microelectromechanical system. The neural network exploits the nonlinear dynamics of a microscale silicon beam to perform its calculations. The group's work looks to create devices that can act simultaneously as a sensor and a computer using a fraction of the energy a normal computer would use.

Reservoir Computing is a framework for computation that may be viewed as an extension of neural networks. Typically an input signal is fed into a fixed (random) dynamical system called a reservoir and the dynamics of the reservoir map the input to a higher dimension. Then a simple readout mechanism is trained to read the state of the reservoir and map it to the desired output. The main benefit is that training is performed only at the readout stage and the reservoir is fixed. Liquid-state machines and echo state networks are two major types of reservoir computing.

New Technique that reduces training time for Deep Learning Networks by more than 60 percent without sacrificing accuracy, accelerating the development of new artificial intelligence (AI) applications.

Networks (computers) - Human Brain - Internet

Liquid State Machine is a type of reservoir computer that uses a spiking neural network. An LSM consists of a large collection of units (called nodes, or neurons). Each node receives time varying input from external sources (the inputs) as well as from other nodes. Nodes are randomly connected to each other. The recurrent nature of the connections turns the time varying input into a spatio-temporal pattern of activations in the network nodes. The spatio-temporal patterns of activation are read out by linear discriminant units. The soup of recurrently connected nodes will end up computing a large variety of nonlinear functions on the input. Given a large enough variety of such nonlinear functions, it is theoretically possible to obtain linear combinations (using the read out units) to perform whatever mathematical operation is needed to perform a certain task, such as speech recognition or computer vision. The word liquid in the name comes from the analogy drawn to dropping a stone into a still body of water or other liquid. The falling stone will generate ripples in the liquid. The input (motion of the falling stone) has been converted into a spatio-temporal pattern of liquid displacement (ripples). LSMs have been put forward as a way to explain the operation of brains. LSMs are argued to be an improvement over the theory of artificial neural networks because: Circuits are not hard coded to perform a specific task. Continuous time inputs are handled "naturally". Computations on various time scales can be done using the same network. The same network can perform multiple computations. Criticisms of LSMs as used in computational neuroscience are that LSMs don't actually explain how the brain functions. At best they can replicate some parts of brain functionality. There is no guaranteed way to dissect a working network and figure out how or what computations are being performed. Very little control over the process.

"Liquid" machine-learning system adapts to changing conditions. MIT researchers have developed a type of neural network that learns on the job, not just during its training phase. These flexible algorithms, dubbed "liquid" networks, change their underlying equations to continuously adapt to new data inputs. The advance could aid decision making based on data streams that change over time, including those involved in medical diagnosis and autonomous driving.

Geoffrey Hinton is an English Canadian cognitive psychologist and computer scientist, most noted for his work on artificial neural networks. Since 2013 he divides his time working for Google (Google Brain) and the University of Toronto. In 2017, he cofounded and became the Chief Scientific Advisor of the Vector Institute in Toronto.

A neural network learns when it should not be trusted. A faster way to estimate uncertainty in AI-assisted decision-making could lead to safer outcomes. Researchers have developed a way for deep learning neural networks to rapidly estimate confidence levels in their output. The advance could enhance safety and efficiency in AI-assisted decision making, with applications ranging from medical diagnosis to autonomous driving. Increasingly, artificial intelligence systems known as deep learning neural networks are used to inform decisions vital to human health and safety, such as in autonomous driving or medical diagnosis. These networks are good at recognizing patterns in large, complex datasets to aid in decision-making. But how do we know they're correct?



Controls


Control is the activity of managing or handling something carefully. Having the power or the ability to direct or determine outcomes. The discipline to regulate functions, actions or reflex's. Control also means to have great skillfulness and having a firm understanding or knowledge of some subject or activity. Control in science is a standard against which other conditions can be compared and verified in a scientific experiment. Top-down and bottom-up.

Adaptive Control is the control method used by a controller which must adapt to a controlled system with parameters which vary, or are initially uncertain.

Process Control in continuous production processes is a combination of control engineering and chemical engineering disciplines that uses industrial control systems to achieve a production level of consistency, economy and safety which could not be achieved purely by human manual control. It is implemented widely in industries such as oil refining, pulp and paper manufacturing, chemical processing and power generating plants, There is a wide range of size, type and complexity, but it enables a small number of operators to manage complex processes to a high degree of consistency. The development of large automatic process control systems was instrumental in enabling the design of large high volume and complex processes, which could not be otherwise economically or safely operated. In process control, there is process gain. Process gain is the relationship between the process control output and the process control input, and is defined as the change in input divided by the change in output. Positive gain is when both the input and the output are increasing, while negative gain is when the input increases, while the output decreases. The applications can range from controlling the temperature and level of a single process vessel, to a complete chemical processing plant with several thousand control loops. Process Control is an engineering discipline that deals with architectures, mechanisms and algorithms for maintaining the output of a specific process within a desired range. For instance, the temperature of a chemical reactor may be controlled to maintain a consistent product output.

Quality Control - Remote Control (the illusion of control)

Manual Control is a type of control that is manually performed by hand or by individuals. Manual is something requiring human control by hand. Backup Manual Control System or Mechanical Backup System provides a person with the ability to operate a machine in the absence of automatic control features.

Thermal Actuator is a device used to transform energy into motion. A thermal actuator is a type of non-electric motor made of components such as a piston and a thermal sensitive material capable of producing linear motion in response to temperature changes. Actuator is a component of a machine that is responsible for moving and controlling a mechanism or system, for example by opening a valve. In simple terms, it is a "mover". Triggers.

Solenoid converts electrical energy into mechanical work. A transducer device that converts energy into linear motion. The term is also often used to refer to a solenoid valve, an integrated device containing an electromechanical solenoid which actuates either a pneumatic or hydraulic valve, or a solenoid switch, which is a specific type of relay that internally uses an electromechanical solenoid to operate an electrical switch.

Electromechanics is the interaction of electrical and mechanical systems as a whole and how the two systems interact with each other. Electromechanical devices are ones which have both electrical and mechanical processes. A manually operated switch is an electromechanical component due to the mechanical movement causing an electrical output.

Harness is to control and direct things as if by using reins. To exploit the power of some force. A harness can also mean a support consisting of an arrangement of straps for holding something to the body, like a parachute. Stable gear consisting of an arrangement of leather straps fitted to a draft animal so that it can be attached to a cart and help pull a cart and be controlled using reins or a pair of long straps usually connected to the headpiece.

Override is to use one's authority to reject or cancel or interrupt the action of an automatic device, typically in order to take manual control and do something more important. You can override or reject a decision. Prevail - Counteract.

Overwrite is entering new data in the place of old data. Editing - Reprogram - Method Overriding.

Nullify is to cancel out, null, void or invalidate a previous decision. Dismiss.

Veto is a constitutional right to reject a decision or proposal made by a law-making body. A veto is the power to unilaterally stop an official action, especially the enactment of legislation. Appeal.

Control Logic is a key part of a software program that controls the operations of the program. The control logic responds to commands from the user, and it also acts on its own to perform automated tasks that have been structured into the program. Control logic can be modeled using a state diagram, which is a form of hierarchical state machine. These state diagrams can also be combined with flow charts to provide a set of computational semantics for describing complex control logic. This mix of state diagrams and flow charts is illustrated in the figure on the right, which shows the control logic for a simple stopwatch. The control logic takes in commands from the user, as represented by the event named “START”, but also has automatic recurring sample time events, as represented by the event named “TIC”.

Control Engineering is an engineering discipline that applies automatic control theory to design systems with desired behaviors in control environments. The discipline of controls overlaps and is usually taught along with electrical engineering at many institutions around the world. The practice uses sensors and detectors to measure the output performance of the process being controlled; these measurements are used to provide corrective feedback helping to achieve the desired performance. Systems designed to perform without requiring human input are called automatic control systems (such as cruise control for regulating the speed of a car). Multi-disciplinary in nature, control systems engineering activities focus on implementation of control systems mainly derived by mathematical modeling of a diverse range of systems. PDF.

Master / Slave Technology is a model of communication where one device or process has unidirectional control over one or more other devices. In some systems a master is selected from a group of eligible devices, with the other devices acting in the role of slaves. In the context of motor control, the master/slave configuration is used for load sharing purposes when two identical motors connected to two different drives are coupled to a common load. One drive is defined as the master and is configured for running in the speed-control mode whereas the other defined as slave is configured for running in torque-control mode.

Control in management is one of the managerial functions like planning, organizing, staffing and directing. It is an important function because it helps to check the errors and to take the corrective action so that deviation from standards are minimized and stated goals of the organization are achieved in a desired manner. According to modern concepts, control is a foreseeing action whereas earlier concept of control was used only when errors were detected. Control in management means setting standards, measuring actual performance and taking corrective action.

You Can't Control Everything - You Can't be Aware of Everything

Possession is the act of having and controlling something tangible or intangible.

Control System is a device, or set of devices, that manages, commands, directs or regulates the behaviour of other devices or systems. They can range from a home heating controller using a thermostat controlling a boiler to large Industrial control systems which are used for controlling processes or machines. Flight Control.

Regulator in automatic control is a regulator is a device which has the function of maintaining a designated characteristic. It performs the activity of managing or maintaining a range of values in a machine. The measurable property of a device is managed closely by specified conditions or an advance set value; or it can be a variable according to a predetermined arrangement scheme. It can be used generally to connote any set of various controls or devices for regulating or controlling items or objects. Examples are a voltage regulator (which can be a transformer whose voltage ratio of transformation can be adjusted, or an electronic circuit that produces a defined voltage), a pressure regulator, such as a diving regulator, which maintains its output at a fixed pressure lower than its input, and a fuel regulator (which controls the supply of fuel). Regulators can be designed to control anything from gases or fluids, to light or electricity. Speed can be regulated by electronic, mechanical, or electro-mechanical means. Such instances include; Electronic regulators as used in modern railway sets where the voltage is raised or lowered to control the speed of the engine. Mechanical Systems such as valves as used in fluid control systems. Purely mechanical pre-automotive systems included such designs as the Watt centrifugal governor whereas modern systems may have electronic fluid speed sensing components directing solenoids to set the valve to the desired rate. Complex electro-mechanical speed control systems used to maintain speeds in modern cars (cruise control) - often including hydraulic components, An aircraft engine's constant speed unit changes the propeller pitch to maintain engine speed. Cybernetics.

Real-time Control System is a reference model architecture, suitable for many software-intensive, real-time control problem domains. RCS is a reference model architecture that defines the types of functions that are required in a real-time intelligent control system, and how these functions are related to each other. Operating Systems.

Programmable Logic Controller is an industrial digital computer which has been ruggedised and adapted for the control of manufacturing processes, such as assembly lines, or robotic devices, or any activity that requires high reliability control and ease of programming and process fault diagnosis. PLCs can range from small modular devices with tens of inputs and outputs (I/O), in a housing integral with the processor, to large rack-mounted modular devices with a count of thousands of I/O, and which are often networked to other PLC and SCADA systems. They can be designed for multiple arrangements of digital and analog I/O, extended temperature ranges, immunity to electrical noise, and resistance to vibration and impact. Programs to control machine operation are typically stored in battery-backed-up or non-volatile memory. PLCs were first developed in the automobile industry to provide flexible, ruggedised and easily programmable controllers to replace hard-wired relays and timers. Since then they have been widely adopted as high-reliability automation controllers suitable for harsh environments. A PLC is an example of a "hard" real-time system since output results must be produced in response to input conditions within a limited time, otherwise unintended operation will result. Algorithms - Electric Motors.

Controller in control theory is a device, historically using mechanical, hydraulic, pneumatic or electronic techniques often in combination, but more recently in the form of a microprocessor or computer, which monitors and physically alters the operating conditions of a given dynamical system. Typical applications of controllers are to hold settings for temperature, pressure, flow or speed.

Nonlinear Control is the area of control theory which deals with systems that are nonlinear, time-variant, or both.

Closed-Loop Transfer Function in control theory is a mathematical expression (algorithm) describing the net result of the effects of a closed (feedback) loop on the input signal to the circuits enclosed by the loop.

Hierarchical Control System is a form of control system in which a set of devices and governing software is arranged in a hierarchical tree. When the links in the tree are implemented by a computer network, then that hierarchical control system is also a form of networked control system.

Intelligent Control is a class of control techniques that use various artificial intelligence computing approaches like neural networks, Bayesian probability, fuzzy logic, machine learning, evolutionary computation and genetic Algorithms.

Networked Control System is a control system wherein the control loops are closed through a communication network. The defining feature of an NCS is that control and feedback signals are exchanged among the system's components in the form of information packages through a network.

Open-Loop Controller is when the control action from the controller is independent of the "process output", which is the process variable that is being controlled. It does not use feedback to determine if its output has achieved the desired goal of the input or process "set point". An open-loop system cannot engage in machine learning and also cannot correct any errors that it could make. It will not compensate for disturbances in the process being controlled.

Perceptual Control Theory is a model of behavior based on the principles of negative feedback, but differing in important respects from engineering control theory. Results of PCT experiments have demonstrated that an organism controls neither its own behavior, nor external environmental variables, but rather its own perceptions of those variables. Actions are not controlled, they are varied so as to cancel the effects that unpredictable environmental disturbances would otherwise have on controlled perceptions.

Automatic Control is the application of mechanisms to the operation and regulation of processes without continuous direct human intervention.

Autonomous - Automation

Control Theory is the idea that two control systems—inner controls and outer controls—work against our tendencies to deviate.

Operating System - Algorithms

Signal Chain is a term used in signal processing and mixed-signal system design to describe a series of signal-conditioning electronic components that receive input (data acquired from sampling either real-time phenomena or from stored data) in tandem, with the output of one portion of the chain supplying input to the next. Signal chains are often used in signal processing applications to gather and process data or to apply system controls based on analysis of real-time phenomena.

Feed Forward in Control is a term describing an element or pathway within a control system which passes a controlling signal from a source in its external environment, often a command signal from an external operator, to a load elsewhere in its external environment. A control system which has only feed-forward behavior responds to its control signal in a pre-defined way without responding to how the load reacts; it is in contrast with a system that also has feedback, which adjusts the output to take account of how it affects the load, and how the load itself may vary unpredictably; the load is considered to belong to the external environment of the system.

Feedback (Positive and Negative) - Placebos

Nothing is beyond your control, there is nothing that you cannot control. Something's are harder to control then others, and there are some things you have not yet learned how to control. To say that I cannot control something is a false statement. To be more accurate, you have to say that I have not yet learned how to control this. Gratification.


Autonomous - Automation


Autopilot is a computer navigation system used to control the trajectory of a vehicle without constant hands-on control by a human operator being required. Autopilots do not replace a human operator, but assist them in controlling the vehicle, allowing them to focus on broader aspects of operation, such as monitoring the trajectory, weather and systems. Autopilots or self-steering gear are used in aircraft, boats, spacecraft, missiles, and others. Autopilots have evolved significantly over time, from early autopilots that merely held an attitude to modern autopilots capable of performing automated landings under the supervision of a pilot. Augment.

Autonomous Robot is a robot that performs behaviors or tasks with a high degree of autonomy, which is particularly desirable in fields such as spaceflight, household maintenance (such as cleaning), waste water treatment and delivering goods and services. Robot Operating System.

Autonomous is something that is not controlled by outside forces. Existing as an independent entity. Free from external control and constraint in e.g. action and judgment. Autonomy is one who gives oneself one's own law and has free-will.

Autonomic Nervous System (autonomous functions of the human body)

Automata Theory is the study of abstract machines and automata, as well as the computational problems that can be solved using them. It is a theory in theoretical computer science, under discrete mathematics (a subject of study in both mathematics and computer science). The word automata (the plural of automaton) comes from the Greek word αὐτόματα, which means "self-acting".

Insects, plants, animals and even matter has built in automation features. Humans have instincts which are extremely useful when knowledge is absent. So automation is necessary. But you still need manual control.

Automation is the use of various control systems for operating equipment such as machinery, processes in factories, boilers and heat treating ovens, switching on telephone networks, steering and stabilization of ships, aircraft and other applications and vehicles with minimal or reduced human intervention. Some processes have been completely automated.

Automaton is a self-operating machine, or a machine or control mechanism designed to follow automatically a predetermined sequence of operations, or respond to predetermined instructions. Some automata, such as bellstrikers in mechanical clocks, are designed to give the illusion to the casual observer that they are operating under their own power. (automata or automatons).

Automation Paradox (off loading) - GMU Autonomous Robotics Laboratory

Automation is Replacing Jobs, so Human Labor will do other more important things, and that's a good thing. There is already Autonomous Machines in Nature, like insects, plants, bacteria, DNA. But these types of autonomous abilities have been perfected over millions of years, and we are just learning how to expand these autonomous abilities to machines. So we need to go slow and learn from the experts in nature, because just like invasive species, autonomous abilities can have catastrophic consequences.

Actuator is a mechanism that puts something into automatic action.

Cam is a rotating or sliding piece in a mechanical linkage used especially in transforming rotary motion into linear motion. It is often a part of a rotating wheel (e.g. an eccentric wheel) or shaft (e.g. a cylinder with an irregular shape) that strikes a lever at one or more points on its circular path. The cam can be a simple tooth, as is used to deliver pulses of power to a steam hammer, for example, or an eccentric disc or other shape that produces a smooth reciprocating (back and forth) motion in the follower, which is a lever making contact with the cam.

Self-Management in computer science is the process by which computer systems shall manage their own operation without human intervention. Self-Management technologies are expected to pervade the next generation of network management systems. The growing complexity of modern networked computer systems is currently the biggest limiting factor in their expansion. The increasing heterogeneity of big corporate computer systems, the inclusion of mobile computing devices, and the combination of different networking technologies like WLAN, cellular phone networks, and mobile ad hoc networks make the conventional, manual management very difficult, time-consuming, and error-prone. More recently self-management has been suggested as a solution to increasing complexity in cloud computing. Currently, the most important industrial initiative towards realizing self-management is the Autonomic Computing Initiative (ACI) started by IBM in 2001. The ACI defines the following four functional areas: Self-Configuration: Automatic configuration of components; Self-Healing: Automatic discovery, and correction of faults; automatically applying all necessary actions to bring system back to normal operation. Self-Optimization: Automatic monitoring and control of resources to ensure the optimal functioning with respect to the defined requirements; Self-Protection: Proactive identification and protection from arbitrary attacks. The design complexity of Autonomic Systems and self-management systems can be simplified by utilizing design patterns such as the Model View Controller (MVC) to improve separation of concerns by helping encapsulate functional concerns.

Impulsivity (lack of control) - Auto-Pilot - Self Driving - Subconscious

Unconscious Mind consists of the processes in the mind which occur automatically and are not available to introspection, and include thought processes, memories, interests, and motivations. Focus.

Group Thinking (influence) - Software (computers) - Smart Home

Nothing is totally Autonomous, nothing is totally independent, nothing is totally free from external control. Nothing is. So what are you talking about when you say something is autonomous?

Everything is Connected - Cause and Effect

Ai is about making humans more effective, it's not about making machines more like humans, because that's crazy. Humans are mistake prone, and machines are supposed to help us reduce mistakes, and help us to analyze our options. A machine could never be more intelligent than the most intelligent human. But a machine could easily be more intelligent than a human who has never learned enough or went to school that did not teach enough. You really don't want a machine to be more intelligent than you, because that clearly says that you don't have the necessary knowledge and information that's needed to be intelligent. But Ai could easily be a teacher and a measuring tool for intelligence, with an emphasis on the word 'Tool'. Ai is not human, or will it ever be. But Ai is a great example and a symbol of human ingenuity and intelligence. A dog is a mans best friend, and Ai is an extension of our friendship, and not a replacement for friendship, for that would be like being friends with yourself. Not exciting or real. But still better then nothing. You can love a machine, but what you are really doing is just loving yourself. A machine could never be a replacement for a human, machine can only be an aid. If we never improve education, or if we keep denying people access to valuable knowledge and information, then yes a machine could be more intelligent than a human who is not fully educated. Ai will not be more intelligent than humans, but Ai will help humans become more intelligent. Ai is the path that we are taking to human intelligence.

Humans are is a sense already a machine, a machine that can create more machines. Machines are not made to replace humans, machines only replace certain actions that humans don't need to do. Thus freeing up humans to do more important work, and also freeing up more time to explore, with more time to relax. Ai advancements will eventually lead us right back to ourselves. There is no better machine then a human. Yes there will be certain machines that will have better abilities in certain areas, but only because we made it so. This way we can focus on other things that are more important. 

Can’t you see, the smarter you make the machine the smarter you become. You say you are going to make intelligent machines, or AI, but on the contrary, it will be the machines that will make you intelligent. And the computer machine has already been doing this for some time. Intelligent machines are just mimics, mirrors, extensions and expansions of the human mind. This is way beyond a paradigm shift. It’s self-realization and enlightenment on the grandest scale. Can’t you see, you are not just building a better machine you are building a better human. And yes not everyone is benefiting from this as fast as we would like, but they will if everyone has a computer and understands what it resembles and what it can achieve. Man is the Machine. And we know how to duplicate this intelligent machine, it's called childbirth plus education. We now have more words and more ways to express them then ever before. Words have the ability to shape the human mind. Words are the machine code or natural language of the brain where they are translated into zero’s and ones so that the synapse knows when to fire and when to create more connections and more associations. We will soon be able to scientifically prove what the correct words should be and when the correct time and sequence they should be learned.

The Lateness of the Hour - Twilight Zone (youtube) - Robots built by Dr. Loren are complete with programmed memories and personalities. The Lateness of the Hour is episode 44 of the American television anthology series The Twilight Zone. It originally aired on December 2, 1960 on CBS.

The human brain is the ferrari of brains. Or you can say that the human brain is the lamborghini of all brains. And from our incredible brains we have created incredible machines as our tools. Tools that makes our brains even more powerful by expanding our abilities. And these tools also save us time, which gives us more time to play, and more time to create more time.


Drones - Sub-Autonomous - Fully Autonomous


Lethal Autonomous Weapon are a type of autonomous military robot that can independently search for and engage targets based on programmed constraints and descriptions. LAW are also called lethal autonomous weapon systems or LAWS, or lethal autonomous robots or LAR, or robotic weapons, or killer robots. LAWs may operate in the air, on land, on water, under water, or in space. The autonomy of current systems as of 2018 was restricted in the sense that a human gives the final command to attack - though there are exceptions with certain "defensive" systems.

People fear autonomous killing robots for a good reason, because we already have programed robots, they're called corrupt politicians, their called narrow minded voters, they're called ignorant people who don't know any better, they're called drug addicts, they're called mindless consumers, they're called anyone who does things just for money, they're called anyone who blindly follows orders, whether internally or externally, or blindly follows the rule of a law without question, they're called police, they're called soldiers, they're called the CIA, they're called the NSA, they're called the IRS, they're called the TSA,  Yes we need a command hierarchy, especially when we have to organize for emergency response, like an incident command system. But when people say "I'm just following orders", what they are really saying is that  can't think for themselves and they have no intelligent reasoning that would allow them to make intelligent decisions on their own. When people blindly follow orders, they act just like a robot. Humans are born free thinkers, but when people are not allowed to think freely for themselves, they are no more than autonomous killing machines. People who have power are also autonomous robots, they have been programmed to hold on to their power. So don't worry about machines killing you, because autonomous humans have killed millions of people, and will continue to kill millions, unless we improve education and improve the media. So until everyone becomes intelligent, this ignorance will continue to Crush, Kill and Destroy (Lost in Space, youtube). You don't want a AI system that makes up its own reasoning or makes up it's own purpose, we already have that, their called humans. We want a machine that thinks, and not a machine that pretends to think. We want a machine to have human qualities, but those qualities need to be defined. You just can't tell a machine to be nice, you need to explain what being nice is, and you need to explain what being good is, and you need to explain what being wrong is, just like what we are supposed to do with humans, which is to educate them.

Westworld (8/10) Movie CLIP - Draw (1973) (youtube) - When Robots can't be controlled. Westworld was a1973 American science-fiction Western thriller film where amusement park androids malfunction and begin killing visitors.

I don't fear Artificial Intelligence, I fear the lack of Intelligence, because ignorance is clearly doing all the damage.

The danger is not Artificial Intelligence, the danger is peoples ignorance. Criminals in power have been using technology to kill for hundreds of years, and not just with drones. When crazy people make machines that can kill humans, that's not artificial intelligence, that's just pure ignorance. Most technologies can be extremely dangerous, especially when technology is used by ignorant people or by criminals. This is another great reason why improving education is a must. When people are more educated on how to use things effectively and efficiently, then these technology abuses will decline and eventually fade away, and the real benefits of technology will be more frequent and more common. An autonomous weapons can be programed to kill, just like some humans can be programmed to kill. But machines have no conscience, which makes it easier for corporations to use machines to commit murder. Machines also don't need sleep, food, water or money. Machines also don't need to be rescued or taken care of when they get injured. Machines also never say no. That is why criminals love machines. Algorithms.

Three Laws of Robotics 1: A Robot may not injure a human being or, through inaction, allow a human being to come to harm. 2: A robot must obey the orders given it by human beings except where such orders would conflict with the First Law. 3: A robot must protect its own existence as long as such protection does not conflict with the First or Second Laws.

It’s nice to have rules, but not everyone follows the rules, or understands the rules. So if you’re going to teach robots how to be smart, I would think that you would also teach humans how to be smart. If you value a machine more than a human, then humans cease to have value. And if humans are not here, then who will care for the machines? The machines?

Principles of Robotics - Human Ethics - Engineering and Physical Sciences Research Council

Artificial Intelligence will not destroy the world, human ignorance will destroy the world, like it has before and is doing now, and that is a fact of life. So what are you doing to rid yourself of ignorance, the same ignorance that has destroyed life in the past and present time? If you don't rid yourself of ignorance, then how will you save the world? How will you save your own life? Will AI save you then? Or will it be your own intelligence that will save you? History has been repeating itself, it's time to break that cycle of failure. It's time for Human Intelligence, because AI will not save us.

Will AI stop politicians from being corrupt? Will AI stop people from graduating from college ignorant and unprepared? Will AI stop people from committing murder? Will AI stop people from committing rape? Will AI stop people from committing child abuse? Will AI stop people from committing theft? Will AI stop people from committing fraud? Will AI stop governments, banks and corporations from starting wars?

War has no future, so there will be no wars in the future. Just like all ignorant behaviors, war will become obsolete and fade away from human life like a bad habit. Humans are not wired for war. War is only a byproduct of the corrupted influences of power. People don't start wars, people in power start wars. Though people are the ones who fight wars, and suffer from the violence from wars, it is the people in power who start wars, and profit from wars. They never fight in wars themselves, for if they did, they would realize how insane and ignorant they are. But sadly, the war machine continues with their propaganda and their story telling fear based narratives that try to manipulate public thinking. War is murder, and murder is illegal. But some how people have been tricked into believing that they are not the same. The war mongers use the media and the movie industries to create war porn and militainment, so as to manipulate people even more. The only way that the war machine lives, is to keep people ignorant. And since ignorance will not be apart of our future, then it's time to let war die.

Meaningful Human Control will only happen when military personnel are educated to be intelligent. In 2011, Air Force psychologists completed a mental-health survey of 600 combat drone operators. Forty-two percent of drone crews reported moderate to high stress, and 20 percent reported emotional exhaustion or burnout. The study’s authors attributed their dire results, in part, to “existential conflict.” A later study found that drone operators suffered from the same levels of depression, anxiety, PTSD, alcohol abuse, and suicidal ideation as traditional combat aircrews. And this is not just about drones, there's long range missile's, large canons and land mines that kill from a distance. Emotionally detached and disconnected from reality.

Tracking a Radio Signal from a Drone is possible, almost in the same way that we track a persons location using their cellphone. Federal Aviation Administration (FAA) signed an agreement to locate operators of drones flying illegally near airports. This technology will allow government to track radio signals of drones within 5-mile radius and find location of operators. It is yet to be decided where this technology is going to be integrated. But if you are in an area where there are multiple signals on the 2.4 GHz band that drones use, you aren’t going to be able to distinguish between them everything is on 2.4 GHz. Cordless phones, WiFi routers, microwaves, garage door openers, keyless entry for cars. Need to safeguard drones and robotic cars against cyber attacks.

Autonomy in Weapon Systems (pdf)

Homing Pigeon was derived from the rock pigeon, selectively bred for its ability to find its way home over extremely long distances. The wild rock pigeon has an innate homing ability, meaning that it will generally return to its nest, (it is believed) using magnetoreception. This made it relatively easy to breed from the birds that repeatedly found their way home over long distances. Flights as long as 1,800 km (1,100 miles) have been recorded by birds in competitive pigeon racing. Their average flying speed over moderate 965 km (600 miles) distances is around 97 km/h (60 miles per hour) and speeds of up to 160 km/h (100 miles per hour) have been observed in top racers for short distances. Because of this skill, homing pigeons were used to carry messages as messenger pigeons.


Self Driving Cars - Along for the Ride


We want machines to have some autonomous abilities, like we do now with operating systems and some cars. But we don't want machines to do things totally on their own. Like, you don't want your computer to shut off or stop running programs when you need them. That is when a human will need the on and off switch, or a cancel button, or the ability to reprogram. Kind of like what we have now with most computers. In order for machines to have intelligent abilities, we first have to have intelligent humans to manage the operation of these intelligent machines. Any type of autonomous Ability in the wrong hands will always have catastrophic consequences, just like we have now, except people are being controlled by money, and not by intelligent algorithms. So we need to focus more on improving the abilities of humans, and focus less on the abilities of machines, or the assumed abilities of machines. We have to understand what having control means.

Self Driving Cars or Vehicular Automation involves the use of mechatronics, artificial intelligence, and multi-agent system to assist a vehicle's operator. These features and the vehicles employing them may be labeled as intelligent or smart. A vehicle using automation for difficult tasks, especially navigation, may be referred to as semi-autonomous. A vehicle relying solely on automation is consequently referred to as robotic or autonomous. After the invention of the integrated circuit, the sophistication of automation technology increased. Manufacturers and researchers subsequently added a variety of automated functions to automobiles and other vehicles. Mobileye software that enables Advanced Driver Assist Systems.

Autonomous Car is unmanned ground vehicle is a vehicle that is capable of sensing its environment and navigating without human input. (also known as a driverless car, self-driving car, robotic car). Cameras.

Advanced Driver-Assistance Systems are systems to help the driver in the driving process. When designed with a safe human-machine interface, they should increase car safety and more generally road safety. Most road accidents occurred due to the human error. Advanced driver-assistance systems are systems developed to automate, adapt and enhance vehicle systems for safety and better driving. The automated system which is provided by ADAS to the vehicle is proven to reduce road fatalities, by minimizing the human error. Safety features are designed to avoid collisions and accidents by offering technologies that alert the driver to potential problems, or to avoid collisions by implementing safeguards and taking over control of the vehicle. Adaptive features may automate lighting, provide adaptive cruise control and collision avoidance, pedestrian crash avoidance mitigation (PCAM), incorporate satnav/traffic warnings, connect to smartphones, alert driver to other cars or dangers, lane departure warning system, automatic lane centering, or show what is in blind spots. Risks.

Level 5 Full Driving Automation. Level 5 cars won't even have steering wheels or acceleration/braking pedals. They will be free from geofencing, able to go anywhere and do anything that an experienced human driver can do.

Geo-Fence is a location-aware device of a location-based service that creates a virtual perimeter area and a predefined set of boundaries or radius around a point location, which can then be used in a real-world geographical area.

Adaptive Cruise Control is an available cruise control system for road vehicles that automatically adjusts the vehicle speed to maintain a safe distance from vehicles ahead.

Telematics is an interdisciplinary field that encompasses telecommunications, vehicular technologies, for instance, road transportation, road safety, electrical engineering (sensors, instrumentation, wireless communications, etc.), and computer science (multimedia, Internet, etc.). Telematics can involve any of the following: The technology of sending, receiving and storing information using telecommunication devices to control remote objects. The integrated use of telecommunications and informatics for application in vehicles and to control vehicles on the move. Global navigation satellite system technology integrated with computers and mobile communications technology in automotive navigation systems. (most narrowly) the use of such systems within road vehicles, also called vehicle telematics. Monitoring Environment.

Drive PX-Series is a series of computers aimed at providing autonomous car and driver assistance functionality powered by deep learning.

Trolley Problem scenario is flawed, incomplete and to general. This is more about determining how ignorant people are then it is trying to determine the ethics of a machine, like with self driving cars. This is like asking someone, "if you were an idiot what would you do?" Since a person could learn nothing from this, then there is no point to these types of thought experiments except to waste time, money, people, resources and so on. The data is almost useless unless you are measuring the level of peoples ignorance. You need to show an actual scenario based on facts and current standards, along with the mechanical limitations and the laws of physics. Then we can determine the choices and options that we have for that particular scenario. Just giving people a choice about something they know very little about, like when people vote in politics, then you have lots of errors with very little understanding of the problems. So in order to accurately measure something, you need to use an example based on reality, and not just a ' what if ' that has many unknown variables. The bottom line is, people make mistakes, which means that algorithms and machines can also make mistakes. And the only way that you can limit your mistakes is by understanding them, which means that you have to know the facts. Learning needs to be the goal of any experiment. Self-driving cars may soon be able to make moral and ethical decisions as humans do, but only when human know better of course.

Driverless cars can actually help teach people how to drive with better awareness. We could use the software that controls the autonomous vehicle, and create a simulation that anyone can use on a computer. It would give people different scenarios that can test a persons awareness. It will make Driving safer and save lives. New Tesla Cars can now make 12 trillion operations a second, almost as good as a Human Brain. And driverless cars are less prone to accidents then a human driver.

Teaching cars to drive with foresight. Good drivers anticipate dangerous situations and adjust their driving before things get dicey. Researchers now also want to teach this skill to self-driving cars. Self-Learning Process.

Enabling autonomous vehicles to see around corners. By sensing tiny changes in shadows, a new system identifies approaching objects that may cause a collision. To improve the safety of autonomous systems, MIT engineers have developed a system that can sense tiny changes in shadows on the ground to determine if there's a moving object coming around the corner.



Machine Learning - Deep Learning


Machine Learning is the study of pattern recognition and computational learning theory in artificial intelligence. ML is the field of study that gives computers the ability to learn without being explicitly programmed. Machine Learning is the construction of algorithms that can learn from and make predictions on data – such algorithms overcome following strictly static program instructions by making data-driven predictions or decisions, through building a model from sample inputs. Outline of Machine Learning (wiki) - PDF - List of Machine Learning Algorithms - Problem Solving - Game Learning.

When someone says that a machine is learning on its own, they mean that they don't have to tell the machine what to do because the programmer already told the machine what to do. It's only when the machine makes a mistake does the human need to interact with the machine again. And that is when the human is training the machine, which comes from human learning, and not from machine learning. So in order for a machine to learn, it still needs to be programmed to learn using specific algorithms and goals. A machine can not just learn on its own like a human does. A machine can only be programmed to follow specific instructions, and that's it. An AI machine has no consciousness, no free will and no soul. And AI machines are not without problems or risk, mostly because humans need to program them, so human error is still a possibility. And autonomy without a conscience is what a criminal is. So don't let criminals make AI machines.

Deep Learning is a branch of machine learning based on a set of algorithms that attempt to model high-level abstractions in data by using a deep graph with multiple processing layers, composed of multiple linear and non-linear transformations. Deep learning uses multiple hidden layers and pooling techniques. Deep learning involves multiple levels of representation and multiple layers of non-linear processing units (or neurons). Deep learning architecture can learn representations and features directly from the input with little to no prior knowledge. Representation learning or hierarchical learning or features learning. Shallow Learning is typically uses only one hidden layer or single layer feed forward networks. Features can be learned more-or-less independently.

Scaling Deep Learning Algorithm leverages Titan to create high-performing deep neural networks - Networks

Meta Learning is a subfield of machine learning where automatic learning algorithms are applied on meta-data about machine learning experiments. Meta Training.

Transfer Learning or inductive transfer is a research problem in machine learning that focuses on storing knowledge gained while solving one problem and applying it to a different but related problem. For example, knowledge gained while learning to recognize cars could apply when trying to recognize trucks. This area of research bears some relation to the long history of psychological literature on transfer of learning, although formal ties between the two fields are limited.

Computational Learning Theory is a subfield of Artificial Intelligence devoted to studying the design and analysis of machine learning. Knowledge-Based System.

Computational Theory of Mind is a view that the human mind or the human brain (or both) is an information processing system and that thinking is a form of computing.

Computational Neuroscience studies brain function in terms of the information processing properties of the structures that make up the nervous system. It is an interdisciplinary computational science that links the diverse fields of neuroscience, cognitive science, and psychology with electrical engineering, computer science, mathematics, and physics.

Computational Model is a mathematical model in computational science that requires extensive computational resources to study the behavior of a complex system by computer simulation.

Computational Complexity Theory is a branch of the theory of computation in theoretical computer science that focuses on classifying computational problems according to their inherent difficulty, and relating those classes to each other. A computational problem is understood to be a task that is in principle amenable to being solved by a computer, which is equivalent to stating that the problem may be solved by mechanical application of mathematical steps, such as an algorithm.

Neuromorphic Engineering describes the use of very-large-scale integration (VLSI) systems containing electronic analog circuits to mimic neuro-biological architectures present in the nervous system. In recent times the term neuromorphic has been used to describe analog, digital, mixed-mode analog/digital VLSI, and software systems that implement models of neural systems (for perception, motor control, or multisensory integration). The implementation of neuromorphic computing on the hardware level can be realized by oxide-based memristors, threshold switches, and transistors. A key aspect of neuromorphic engineering is understanding how the morphology of individual neurons, circuits, applications, and overall architectures creates desirable computations, affects how information is represented, influences robustness to damage, incorporates learning and development, adapts to local change (plasticity), and facilitates evolutionary change.

Cognitive Computer combines artificial intelligence and machine-learning algorithms, in an approach which attempts to reproduce the behaviour of the human brain.

Evolutionary Computation is a family of algorithms for global optimization inspired by biological evolution, and the subfield of artificial intelligence and soft computing studying these algorithms. In technical terms, they are a family of population-based trial and error problem solvers with a metaheuristic or stochastic optimization character.

Computer Code - Super Computers

Algorithm that Learns directly from Human Instructions, rather than an existing set of examples, and outperformed conventional methods of training neural networks by 160 per cent.

Zero Shot Learning is a form of extending supervised learning to a setting of solving for example a classification problem when there's not enough labeled examples available for all classes, or there is not enough training data available. If you never seen that animal before, how could you guess the name of the animal?

Hierarchical Temporal Memory are learning algorithms that can store, learn, infer and recall high-order sequences. Unlike most other machine learning methods, HTM learns (in an unsupervised fashion) time-based patterns in unlabeled data on a continuous basis. HTM is robust to noise, and it has high capacity, meaning that it can learn multiple patterns simultaneously. When applied to computers, HTM is well suited for prediction, anomaly detection, classification and ultimately sensorimotor applications. The theory has been tested and implemented in software through example applications from Numenta and a few commercial applications from Numenta's partners. Cognitive Hierarchy.

With little training, machine-learning algorithms can uncover hidden scientific knowledge. Researchers have shown that an Algorithm with no training in materials science can scan the text of millions of papers and uncover new scientific knowledge. They collected 3.3 million abstracts of published materials science papers and fed them into an algorithm called Word2vec. By analyzing relationships between words the algorithm was able to predict discoveries of new thermoelectric materials years in advance and suggest as-yet unknown materials as candidates for thermoelectric materials.

Robot Learning (PDF) - Robotics - Sensors

Tweaking AI software to function like a human brain improves computer's learning ability. Humans can quickly and accurately learn new visual concepts from sparse data ¬- sometimes just a single example. Even three- to four-month-old babies can easily learn to recognize zebras and distinguish them from cats, horses, and giraffes. But computers typically need to "see" many examples of the same object to know what it is.

Teachable Machine Experiment using your camera, live in the browser. No coding required. (google)

Internet of Things - Learning Objectives (purpose)

Machine Learning is just Human Learning using a Machine. Machine learning is more about Human Learning. It's humans learning what they want machines to do and then recording that knowledge into a machine. Then humans program the machine so it knows how to interpret that knowledge effectively and efficiently. That's what they are supposed to be doing, anyway. Algorithm is calculations and formulas that we choose to use that will give us the answers that we are looking for, and when the machine gets the right answer, then the algorithm works for that type of problem solving.

Machine Learning is trying to do what DNA has been doing for millions of years, make the best decisions possible using past knowledge along with the current information acquired from the environment. The goal of all life is to adapt, create balance, reduce vulnerabilities and ultimately survive. Learning is key. Define the inputs, define the desired outputs, and pay attention to any unusual changes that happen, changes that would require a modification to the inputs or to the outputs. Document. The reward is a measured improvement that created more stability and a better quality of living. The system will always keep looking for a way to make another improvement and receive another reward. Even when things get bad, it will only mean that there is now more room for improvements, thus, more rewards to receive. Intelligence Formula.

Prior Knowledge for Pattern Recognition refers to all information about the problem available in addition to the training data. However, in this most general form, determining a model from a finite set of samples without prior knowledge is an ill-posed problem, in the sense that a unique model may not exist. Many classifiers incorporate the general smoothness assumption that a test pattern similar to one of the training samples tends to be assigned to the same class. The importance of prior knowledge in machine learning is suggested by its role in search and optimization. Loosely, the no free lunch theorem states that all search algorithms have the same average performance over all problems, and thus implies that to gain in performance on a certain application one must use a specialized algorithm that includes some prior knowledge about the problem. The different types of prior knowledge encountered in pattern recognition are now regrouped under two main categories: class-invariance and knowledge on the data. Pattern recognition is a very active field of research intimately bound to machine learning. Also known as classification or statistical classification, pattern recognition aims at building a classifier that can determine the class of an input pattern. This procedure, known as training, corresponds to learning an unknown decision function based only on a set of input-output pairs that form the training data (or training set). Nonetheless, in real world applications such as character recognition, a certain amount of information on the problem is usually known beforehand. The incorporation of this prior knowledge into the training is the key element that will allow an increase of performance in many applications.

Training, Test, and Validation Sets in machine learning, the study and construction of algorithms that can learn from and make predictions on data is a common task. Such algorithms work by making data-driven predictions or decisions, :2 through building a mathematical model from input data. The data used to build the final model usually comes from multiple datasets. In particular, three data sets are commonly used in different stages of the creation of the model. The model is initially fit on a training dataset, that is a set of examples used to fit the parameters (e.g. weights of connections between neurons in artificial neural networks) of the model. The model (e.g. a neural net or a naive Bayes classifier) is trained on the training dataset using a supervised learning method (e.g. gradient descent or stochastic gradient descent). In practice, the training dataset often consist of pairs of an input vector and the corresponding answer vector or scalar, which is commonly denoted as the target. The current model is run with the training dataset and produces a result, which is then compared with the target, for each input vector in the training dataset. Based on the result of the comparison and the specific learning algorithm being used, the parameters of the model are adjusted. The model fitting can include both variable selection and parameter estimation. Successively, the fitted model is used to predict the responses for the observations in a second dataset called the validation dataset. The validation dataset provides an unbiased evaluation of a model fit on the training dataset while tuning the model's hyperparameters (e.g. the number of hidden units in a neural network). Validation datasets can be used for regularization by early stopping: stop training when the error on the validation dataset increases, as this is a sign of overfitting to the training dataset. This simple procedure is complicated in practice by the fact that the validation dataset's error may fluctuate during training, producing multiple local minima. This complication has led to the creation of many ad-hoc rules for deciding when overfitting has truly begun. Finally, the test dataset is a dataset used to provide an unbiased evaluation of a final model fit on the training dataset.

Machine to Machine refers to direct communication between devices using any communications channel, including wired and wireless. Machine to machine communication can include industrial instrumentation, enabling a sensor or meter to communicate the data it records (such as temperature, inventory level, etc.) to application software that can use it (for example, adjusting an industrial process based on temperature or placing orders to replenish inventory). Such communication was originally accomplished by having a remote network of machines relay information back to a central hub for analysis, which would then be rerouted into a system like a personal computer.

Intelligence Amplification refers to the effective use of information technology in augmenting human intelligence.

Computer Vision

Adversarial Machine Learning is a technique employed in the field of machine learning which attempts to fool models through malicious input. This technique can be applied for a variety of reasons, the most common being to attack or cause a malfunction in standard machine learning models. Machine learning techniques were originally designed for stationary and benign environments in which the training and test data are assumed to be generated from the same statistical distribution. However, when those models are implemented in the real world, the presence of intelligent and adaptive adversaries may violate that statistical assumption to some degree, depending on the adversary. This technique shows how a malicious adversary can surreptitiously manipulate the input data so as to exploit specific vulnerabilities of learning algorithms and compromise the security of the machine learning system.

Generative Adversarial Network is when two neural networks contest with each other in a game (in the sense of game theory, often but not always in the form of a zero-sum game). Given a training set, this technique learns to generate new data with the same statistics as the training set. For example, a GAN trained on photographs can generate new photographs that look at least superficially authentic to human observers, having many realistic characteristics. Though originally proposed as a form of generative model for unsupervised learning, GANs have also proven useful for semi-supervised learning, fully supervised learning, and reinforcement learning. In a 2016 seminar, Yann LeCun described GANs as "the coolest idea in machine learning in the last twenty years".

Adversarial Reprogramming of Neural Networks. Deep neural networks are susceptible to adversarial attacks. It enables attackers to fool systems in what are known as “black-box attacks” where they don't have access to the model's architecture, parameters or even the training data used to train the network. In computer vision, well-crafted perturbations to images can cause neural networks to make mistakes such as confusing a cat with a computer. Previous adversarial attacks have been designed to degrade performance of models or cause machine learning models to produce specific outputs chosen ahead of time by the attacker. We introduce attacks that instead reprogram the target model to perform a task chosen by the attacker without the attacker needing to specify or compute the desired output for each test-time input. This attack finds a single adversarial perturbation, that can be added to all test-time inputs to a machine learning model in order to cause the model to perform a task chosen by the adversary—even if the model was not trained to do this task. These perturbations can thus be considered a program for the new task. We demonstrate adversarial reprogramming on six ImageNet classification models, repurposing these models to perform a counting task, as well as classification tasks: classification of MNIST and CIFAR-10 examples presented as inputs to the ImageNet model.

Deep Learning - Deep-Learning Program DRIVE PX

Deep Learning & Artificial Intelligence Solutions from NVIDIA

Weights and Biases is on a mission to build the best software tools for machine learning.

Similarity Learning is an area of supervised machine learning in artificial intelligence. It is closely related to regression and classification, but the goal is to learn from examples a similarity function that measures how similar or related two objects are. It has applications in ranking, in recommendation systems, visual identity tracking, face verification, and speaker verification.

Domain Randomization is a simple technique for training models on simulated images that transfer to real images by randomizing rendering in the simulator. With enough variability in the simulator, the real world may appear to the model as just another variation.

Object Localization is to predict the object in an image as well as its boundaries. The difference between object localization and object detection is subtle. Simply, object localization aims to locate the main (or most visible) object in an image while object detection tries to find out all the objects and their boundaries.

Stages of Learning

Monad Functional Programming are a way to build computer programs by joining simple components in robust ways. Monads can be seen as a functional design pattern to build generic types, with the following organization: Define a data type, and how values of that datatype are combined. Create functions that use the data type, and compose them together (following the rules defined in the first step).

Human Learning Methods

Statistical Learning Theory is a framework for machine learning drawing from the fields of statistics and functional analysis. Statistical learning theory deals with the problem of finding a predictive function based on data. Statistical learning theory has led to successful applications in fields such as computer vision, speech recognition, bioinformatics and baseball.

Learning Games

Linear Algebra is the branch of mathematics concerning vector spaces and linear mappings between such spaces. It includes the study of lines, planes, and subspaces, but is also concerned with properties common to all vector spaces.

Reinforcement Learning differs from standard supervised learning in that correct input/output pairs are never presented, nor sub-optimal actions explicitly corrected. Instead the focus is on on-line performance, which involves finding a balance between exploration (of uncharted territory) and exploitation (of current knowledge). Agent - Praise, Punish and Reward.

Reinforcement Learning Algorithms - TRPO, DQN, A3C, DDPG, DPO, Rainbw

More effective training model for robots. The underlying adaptation and re-planning mechanism consists of reinforcement learning-based policies. Making these policies efficiently obtainable is critical to making the MDO operating concept a reality. Policy gradient methods in reinforcement learning are the foundation for scalable algorithms for continuous spaces, but existing techniques cannot incorporate broader decision-making goals such as risk sensitivity, safety constraints, exploration and divergence to a prior. Designing autonomous behaviors when the relationship between dynamics and goals are complex may be addressed with reinforcement learning. Reducing the volatility of reward accumulation, ensuring one explores an unknown domain in an efficient manner, or incorporating prior experience, all contribute towards breaking existing sample efficiency barriers of prevailing practice in reinforcement learning by alleviating the amount of random sampling one requires in order to complete policy optimization.

Credit-Assignment is the process of identifying among the set of actions chosen in an episode the ones which are responsible for the final outcome. And moreover, it is an attempt to identify the best, and worst, decisions chosen during an episode, so that the best decisions are reinforced and the worst penalized.

Markov Decision Process provide a mathematical framework for modeling decision making in situations where outcomes are partly random and partly under the control of a decision maker.

Unsupervised Learning is the machine learning task of inferring a function to describe hidden structure from unlabeled data. Since the examples given to the learner are unlabeled, there is no error or reward signal to evaluate a potential solution – this distinguishes unsupervised learning from supervised learning and reinforcement learning. Unsupervised learning is closely related to the problem of density estimation in statistics. However, unsupervised learning also encompasses many other techniques that seek to summarize and explain key features of the data. Knowledge.

Supervised Learning is the machine learning task of inferring a function from labeled training data. The training data consist of a set of training examples. In supervised learning, each example is a pair consisting of an input object (typically a vector) and a desired output value (also called the supervisory signal). A supervised learning algorithm analyzes the training data and produces an inferred function, which can be used for mapping new examples. An optimal scenario will allow for the algorithm to correctly determine the class labels for unseen instances. This requires the learning algorithm to generalize from the training data to unseen situations in a "reasonable" way (inductive bias). The parallel task in human and animal psychology is often referred to as concept learning.

Learning Neural Network. A new type of neural network made with memristors can dramatically improve the efficiency of teaching machines to think like humans. The network, called a reservoir computing system, could predict words before they are said during conversation, and help predict future outcomes based on the present.

Memristor is an electrical component that limits or regulates the flow of electrical current in a circuit and remembers the amount of charge that has previously flowed through it. Memristors are important because they are non-volatile, meaning that they retain memory without power. A hypothetical non-linear passive two-terminal electrical component relating electric charge and magnetic flux linkage.

Reservoir Computing is a framework for computation that may be viewed as an extension of neural networks. Typically an input signal is fed into a fixed (random) dynamical system called a reservoir and the dynamics of the reservoir map the input to a higher dimension. Then a simple readout mechanism is trained to read the state of the reservoir and map it to the desired output. The main benefit is that the training is performed only at the readout stage and the reservoir is fixed. Liquid-state machines and echo state networks are two major types of reservoir computing.

Cognitive Model is an approximation to animal cognitive processes (predominantly human) for the purposes of comprehension and prediction. Cognitive models can be developed within or without a cognitive architecture, though the two are not always easily distinguishable.

International Conference on Machine Learning (wiki) - ICML Website

Inference Engine is a component of the system that applies logical rules to the knowledge base to deduce new information. The first inference engines were components of expert systems. The typical expert system consisted of a knowledge base and an inference engine. The knowledge base stored facts about the world. The inference engine applies logical rules to the knowledge base and deduced new knowledge. This process would iterate as each new fact in the knowledge base could trigger additional rules in the inference engine. Inference engines work primarily in one of two modes either special rule or facts: forward chaining and backward chaining. Forward chaining starts with the known facts and asserts new facts. Backward chaining starts with goals, and works backward to determine what facts must be asserted so that the goals can be achieved.

Human Operating System - Teaching Machine - Computer Science

Numenta reverse engineering the neocortex.

Framework improves 'continual learning' for Artificial Intelligence. Researchers have developed a new framework for deep neural networks that allows artificial intelligence (AI) systems to better learn new tasks while 'forgetting' less of what it has learned regarding previous tasks. The researchers have also demonstrated that using the framework to learn a new task can make the AI better at performing previous tasks, a phenomenon called backward transfer. When asking a deep neural network to learn a new task, the Learn to Grow framework begins by conducting something called an explicit neural architecture optimization via search. What this means is that as the network comes to each layer in its system, it can decide to do one of four things: skip the layer; use the layer in the same way that previous tasks used it; attach a lightweight adapter to the layer, which modifies it slightly; or create an entirely new layer. This architecture optimization effectively lays out the best topology, or series of layers, needed to accomplish the new task. Once this is complete, the network uses the new topology to train itself on how to accomplish the task -- just like any other deep learning AI system.

Machine learning predicts behavior of biological circuits. Neural networks cut modeling times of complex biological circuits to enable new insights into their inner workings.

AI machines can solving complex problems just as accurately as scientists, but considerably faster. Scientists used the tensorial kernel to equip a "support vector machine," which is able to categorize complex data into different groups. The Munich scientists fed the machine a quarter of a million spin configurations generated by the OIST supercomputer simulations of the pyrochlore model. Without any information about which phases were present, the machine successfully managed to reproduce an identical version of the phase diagram. Importantly, when the scientists deciphered the "decision function" which the machine had constructed to classify different types of spin liquid, they found that the computer had also independently figured out the exact mathematical equations that exemplified each phase -- with the whole process taking a matter of weeks.

Machine Learning model helps characterize compounds for drug discovery. Innovators have created a new method of applying machine learning concepts to the tandem mass spectrometry process to improve the flow of information in the development of new drugs.

Machine Learning as an Adversarial Service: Learning Black-Box Adversarial Examples. When a group of researchers from Google and OpenAI realized they could slightly shift the pixels in an image so that it would appear the same to the human eye, but a machine learning algorithm would classify it as something else entirely. For instance, an image might look like a cat to you, but when a computer vision program looks at it, it sees a dog.

Quantum Machine Learning is the integration of quantum algorithms within machine learning programs. The most common use of the term refers to machine learning algorithms for the analysis of classical data executed on a quantum computer, i.e. quantum-enhanced machine learning. While machine learning algorithms are used to compute immense quantities of data, quantum machine learning utilizes qubits and quantum operations or specialized quantum systems to improve computational speed and data storage done by algorithms in a program. This includes hybrid methods that involve both classical and quantum processing, where computationally difficult subroutines are outsourced to a quantum device. These routines can be more complex in nature and executed faster on a quantum computer. Furthermore, quantum algorithms can be used to analyze quantum states instead of classical data. Beyond quantum computing, the term "quantum machine learning" is also associated with classical machine learning methods applied to data generated from quantum experiments (i.e. machine learning of quantum systems), such as learning the phase transitions of a quantum system or creating new quantum experiments. Quantum machine learning also extends to a branch of research that explores methodological and structural similarities between certain physical systems and learning systems, in particular neural networks. For example, some mathematical and numerical techniques from quantum physics are applicable to classical deep learning and vice versa. Furthermore, researchers investigate more abstract notions of learning theory with respect to quantum information, sometimes referred to as "quantum learning theory".



Algorithms


Euclid flowchart Algorithm is a self-contained step-by-step set of operations to be performed. Algorithms perform calculation, data processing, and/or automated reasoning tasks. Algorithm is a precise rule or set of rules that specifying how to solve a particular problem.

Procedure - Formula - Patterns - Variables - What If - Scenarios - Decision Table

Task in computing is a unit of execution or a unit of work. The term is ambiguous; precise alternative terms include process, light-weight process, thread (for execution), step, request, or query (for work). In the adjacent diagram, there are queues of incoming work to do and outgoing completed work, and a thread pool of threads to perform this work. Either the work units themselves or the threads that perform the work can be referred to as "tasks", and these can be referred to respectively as requests/responses/threads, incoming tasks/completed tasks/threads (as illustrated), or requests/responses/tasks. Thread in computing is an execution that is the smallest sequence of programmed instructions that can be managed independently by a scheduler, which is typically a part of the operating system.

Time Complexity is the computational complexity that describes the amount of time it takes to run an algorithm. Time complexity is commonly estimated by counting the number of elementary operations performed by the algorithm, supposing that each elementary operation takes a fixed amount of time to perform. Thus, the amount of time taken and the number of elementary operations performed by the algorithm are taken to differ by at most a constant factor.

Human-Based Genetic Algorithm is a genetic algorithm that allows humans to contribute solution suggestions to the evolutionary process. For this purpose, a HBGA has human interfaces for initialization, mutation, and recombinant crossover. As well, it may have interfaces for selective evaluation. In short, a HBGA outsources the operations of a typical genetic algorithm to humans.

Feedback (open loop) - Machine Learning - Reasoning (intelligence)

Genetic Algorithm is a metaheuristic inspired by the process of natural selection that belongs to the larger class of evolutionary algorithms (EA). Genetic algorithms are commonly used to generate high-quality solutions to optimization and search problems by relying on bio-inspired operators such as mutation, crossover and selection.

Algorithms and Applications for answering Ranked Queries using Ranked Views (PDF)

Analysis of Algorithms is the determination of the computational complexity of algorithms, that is the amount of time, storage and/or other resources necessary to execute them.

Computer Algebra is a scientific area that refers to the study and development of algorithms and software for manipulating mathematical expressions and other mathematical objects. Although computer algebra could be considered a subfield of scientific computing, they are generally considered as distinct fields because scientific computing is usually based on numerical computation with approximate floating point numbers, while symbolic computation emphasizes exact computation with expressions containing variables that have no given value and are manipulated as symbols.

Super-Recursive Algorithm are a generalization of ordinary algorithms that are more powerful, that is, compute more than Turing machines. Turing machines and other mathematical models of conventional algorithms allow researchers to find properties of recursive algorithms and their computations. In a similar way, mathematical models of super-recursive algorithms, such as inductive Turing machines, allow researchers to find properties of super-recursive algorithms and their computations.

Perceptron is an algorithm for supervised learning of binary classifiers. A binary classifier is a function which can decide whether or not an input, represented by a vector of numbers, belongs to some specific class. It is a type of linear classifier, i.e. a classification algorithm that makes its predictions based on a linear predictor function combining a set of weights with the feature vector.

Sorting Algorithm is an algorithm that puts elements of a list in a certain order. The most-used orders are numerical order and lexicographical order. Efficient sorting is important for optimizing the use of other algorithms (such as search and merge algorithms) which require input data to be in sorted lists; it is also often useful for canonicalizing data and for producing human-readable output. More formally, the output must satisfy two conditions: The output is in nondecreasing order (each element is no smaller than the previous element according to the desired total order); The output is a permutation (reordering) of the input. Further, the data is often taken to be in an array, which allows random access, rather than a list, which only allows sequential access, though often algorithms can be applied with suitable modification to either type of data.

Counting Sort is an algorithm for sorting a collection of objects according to keys that are small integers; that is, it is an integer sorting algorithm. It operates by counting the number of objects that have each distinct key value, and using arithmetic on those counts to determine the positions of each key value in the output sequence. Its running time is linear in the number of items and the difference between the maximum and minimum key values, so it is only suitable for direct use in situations where the variation in keys is not significantly greater than the number of items. However, it is often used as a subroutine in another sorting algorithm, radix sort, that can handle larger keys more efficiently. Because counting sort uses key values as indexes into an array, it is not a comparison sort, and the O(n log n) lower bound for comparison sorting does not apply to it. Bucket sort may be used for many of the same tasks as counting sort, with a similar time analysis; however, compared to counting sort, bucket sort requires linked lists, dynamic arrays or a large amount of preallocated memory to hold the sets of items within each bucket, whereas counting sort instead stores a single number (the count of items) per bucket.

Selection Algorithm is an algorithm for finding the kth smallest number in a list or array; such a number is called the kth order statistic. This includes the cases of finding the minimum, maximum, and median elements.

Odds Algorithm is a mathematical method for computing optimal strategies for a class of problems that belong to the domain of optimal stopping problems.

Critical Path Method is an algorithm for scheduling a set of project activities.

Turing Machine

Inductive Turing Machines implement an important class of super-recursive algorithms. An inductive Turing Machine is a definite list of well-defined instructions for completing a task which, when given an initial state, will proceed through a well-defined series of successive states, eventually giving the final result. The difference between an inductive Turing machine and an ordinary Turing machine is that an ordinary Turing machine must stop when it has obtained its result, while in some cases an inductive Turing machine can continue to compute after obtaining the result, without stopping.

Turing Machine is an abstract machine that manipulates symbols on a strip of tape according to a table of rules; to be more exact, it is a mathematical model of computation that defines such a device. Despite the model's simplicity, given any computer algorithm, a Turing machine can be constructed that is capable of simulating that algorithm's logic. Turing Test.

Universal Turing Machine essentially achieves this by reading both the description of the machine to be simulated as well as the input to that machine from its own tape. Alan Turing introduced the idea of such a machine in 1936–1937. This principle is considered to be the origin of the idea of a stored-program computer used by John von Neumann in 1946 for the "Electronic Computing Instrument" that now bears von Neumann's name: the von Neumann architecture. In terms of computational complexity, a multi-tape universal Turing machine need only be slower by logarithmic factor compared to the machines it simulates. (one computer can run all kinds of different programs and can be programmed to do almost anything you want).

Computational Universe is proposed by Jürgen Schmidhuber in a paper based on Zuse's 1967 thesis. He pointed out that a simple explanation of the universe would be a Turing machine programmed to execute all possible programs computing all possible histories for all types of computable physical laws. Computational Biology.

Computation and the Fundamental Theory of Physics - with Stephen Wolfram (youtube) - Stephen Wolfram discusses his efforts to use what he's learned from exploring computational systems to build a new fundamental theory of all of physics.

Universal Computation rests on the principle of simulation and is one of the foundational concepts in computer science. Any computation that can be carried out by one general-purpose computer can also be carried out on any other general-purpose computer. Universal.

Turing Completeness is a system of data-manipulation rules (such as a computer's instruction set, a programming language, or a cellular automaton) is said to be Turing-complete or computationally universal if it can be used to simulate any Turing machine. This means that this system is able to recognize or decide other data-manipulation rule sets. Turing completeness is used as a way to express the power of such a data-manipulation rule set. Virtually all programming languages today are Turing-complete. Turing Test.

Computability is the ability to solve a problem in an effective manner. The computability of a problem is closely linked to the existence of an algorithm to solve the problem. Technology Advancement.

Computation is any type of calculation that includes both arithmetical and non-arithmetical steps and which follows a well-defined model (e.g. an algorithm). Mechanical or electronic devices (or, historically, people) that perform computations are known as computers. An especially well-known discipline of the study of computation is computer science.

Computational Problem is a problem that a computer might be able to solve or a question that a computer may be able to answer. A computational problem is a task solved by a computer. A computation problem is solvable by mechanical application of mathematical steps, such as an algorithm. A problem is regarded as inherently difficult if its solution requires significant resources, whatever the algorithm used. Computational Equivalence says that systems found in the natural world can perform computations up to a maximal ("universal") level of computational power, and that most systems do in fact attain this maximal level of computational power.

Computational Complexity Theory focuses on classifying computational problems according to their inherent difficulty, and relating these classes to each other. It formalizes computational problems by introducing mathematical models of computation to study these problems and quantifying their computational complexity, i.e., the amount of resources needed to solve them, such as time and storage. Other measures of complexity are also used, such as the amount of communication (used in communication complexity), the number of gates in a circuit (used in circuit complexity) and the number of processors (used in parallel computing). One of the roles of computational complexity theory is to determine the practical limits on what computers can and cannot do. The P versus NP problem, one of the seven Millennium Prize Problems, is dedicated to the field of computational complexity. Closely related fields in theoretical computer science are analysis of algorithms and computability theory. A key distinction between analysis of algorithms and computational complexity theory is that the former is devoted to analyzing the amount of resources needed by a particular algorithm to solve a problem, whereas the latter asks a more general question about all possible algorithms that could be used to solve the same problem. More precisely, computational complexity theory tries to classify problems that can or cannot be solved with appropriately restricted resources. In turn, imposing restrictions on the available resources is what distinguishes computational complexity from computability theory: the latter theory asks what kinds of problems can, in principle, be solved algorithmically. A computational problem is a task solved by a computer. A computation problem is solvable by mechanical application of mathematical steps, such as an algorithm. A problem is regarded as inherently difficult if its solution requires significant resources, whatever the algorithm used.

Kolmogorov Complexity of an object, such as a piece of text, is the length of a shortest computer program (in a predetermined programming language) that produces the object as output. It is a measure of the computational resources needed to specify the object, and is also known as algorithmic complexity.

Combinatory Logic is a notation to eliminate the need for quantified variables in mathematical logic.

Bombe is an electro-mechanical device used by British cryptologists to help decipher German Enigma-machine-encrypted secret messages during World War II. The US Navy and US Army later produced their own machines to the same functional specification, albeit engineered differently both from each other and from the British Bombe itself.

Emergent Algorithm is an algorithm that exhibits emergent behavior. In essence an emergent algorithm implements a set of simple building block behaviors that when combined exhibit more complex behaviors. One example of this is the implementation of fuzzy motion controllers used to adapt robot movement in response to environmental obstacles. An emergent algorithm has the following characteristics: it achieves predictable global effects, it does not require global visibility, it does not assume any kind of centralized control, it is self-stabilizing. Other examples of emergent algorithms and models include cellular automata, artificial neural networks and swarm intelligence systems (ant colony optimization, bees algorithm, etc.).

Randomized Algorithm is an algorithm that employs a degree of randomness as part of its logic. The algorithm typically uses uniformly random bits as an auxiliary input to guide its behavior, in the hope of achieving good performance in the "average case" over all possible choices of random bits. Formally, the algorithm's performance will be a random variable determined by the random bits; thus either the running time, or the output (or both) are random variables.

Deterministic Algorithm is an algorithm which, given a particular input, will always produce the same output, with the underlying machine always passing through the same sequence of states. Deterministic algorithms are by far the most studied and familiar kind of algorithm, as well as one of the most practical, since they can be run on real machines efficiently. Formally, a deterministic algorithm computes a mathematical function; a function has a unique value for any input in its domain, and the algorithm is a process that produces this particular value as output.

Nondeterministic Algorithm is an algorithm that, even for the same input, can exhibit different behaviors on different runs, as opposed to a deterministic algorithm. There are several ways an algorithm may behave differently from run to run. A concurrent algorithm can perform differently on different runs due to a race condition. A probabilistic algorithm's behaviors depends on a random number generator. An algorithm that solves a problem in nondeterministic polynomial time can run in polynomial time or exponential time depending on the choices it makes during execution. The nondeterministic algorithms are often used to find an approximation to a solution, when the exact solution would be too costly to obtain using a deterministic one.

Algorithmic Learning Theory is a mathematical framework for analyzing machine learning problems and algorithms. Synonyms include formal learning theory and algorithmic inductive inference. Algorithmic learning theory is different from statistical learning theory in that it does not make use of statistical assumptions and analysis. Both algorithmic and statistical learning theory are concerned with machine learning and can thus be viewed as branches of computational learning theory.

Evolutionary Algorithm is a subset of evolutionary computation, a generic population-based metaheuristic optimization algorithm. An EA uses mechanisms inspired by biological evolution, such as reproduction, mutation, recombination, and selection. Candidate solutions to the optimization problem play the role of individuals in a population, and the fitness function determines the quality of the solutions (see also loss function). Evolution of the population then takes place after the repeated application of the above operators. Artificial evolution (AE) describes a process involving individual evolutionary algorithms; EAs are individual components that participate in an AE.

Memetic Algorithm referrs to in the literature as Baldwinian evolutionary algorithms (EAs), Lamarckian EAs, cultural algorithms, or genetic local search. Algorithms.

Expectation Maximization Algorithm is an iterative method to find maximum likelihood or maximum a posteriori (MAP) estimates of parameters in statistical models, where the model depends on unobserved latent variables. The EM iteration alternates between performing an expectation (E) step, which creates a function for the expectation of the log-likelihood evaluated using the current estimate for the parameters, and a maximization (M) step, which computes parameters maximizing the expected log-likelihood found on the E step. These parameter-estimates are then used to determine the distribution of the latent variables in the next E step.

Monad Functional Programming are a way to build computer programs by joining simple components in robust ways. A monad may encapsulate values of a particular data type, creating a new type associated with a specific computation.

Precondition is a condition or predicate that must always be true just prior to the execution of some section of code or before an operation in a formal specification. If a precondition is violated, the effect of the section of code becomes undefined and thus may or may not carry out its intended work. Security problems can arise due to incorrect preconditions. Often, preconditions are simply included in the documentation of the affected section of code. Preconditions are sometimes tested using guards or assertions within the code itself, and some languages have specific syntactic constructions for doing so. For example: the factorial is only defined for integers greater than or equal to zero. So a program that calculates the factorial of an input number would have preconditions that the number be an integer and that it be greater than or equal to zero.

Algorithm Aversion (PDF)

Parallel Algorithm as opposed to a traditional serial algorithm, is an algorithm which can be executed a piece at a time on many different processing devices, and then combined together again at the end to get the correct result. Many parallel algorithms are executed concurrently – though in general concurrent algorithms are a distinct concept – and thus these concepts are often conflated, with which aspect of an algorithm is parallel and which is concurrent not being clearly distinguished. Further, non-parallel, non-concurrent algorithms are often referred to as "sequential algorithms", by contrast with concurrent algorithms.

Errors (lies)

Callback is any executable code that is passed as an argument to other code, which is expected to call back (execute) the argument at a given time. This execution may be immediate as in a synchronous callback, or it might happen at a later time as in an asynchronous callback. In all cases, the intention is to specify a function or subroutine as an entity that is, depending on the language, more or less similar to a variable. Programming languages support callbacks in different ways, often implementing them with subroutines, lambda expressions, blocks, or function pointers.

Controls (programmable controllers) - Patterns (recognition) - Programming (code)

Instance Based Learning Algorithm (PDF)

Bron-Kerbosch Algorithm is an algorithm for finding maximal cliques in an undirected graph. That is, it lists all subsets of vertices with the two properties that each pair of vertices in one of the listed subsets is connected by an edge, and no listed subset can have any additional vertices added to it while preserving its complete connectivity.

Big O Notation is a mathematical notation that describes the limiting behavior of a function when the argument tends towards a particular value or infinity.

Binary Search Algorithm is a search algorithm that finds the position of a target value within a sorted array. Binary search compares the target value to the middle element of the array; if they are unequal, the half in which the target cannot lie is eliminated and the search continues on the remaining half until it is successful or the remaining half is empty.

Algorithmic Probability is a mathematical method of assigning a prior probability to a given observation. It was invented by Ray Solomonoff in the 1960s. It is used in inductive inference theory and analyses of algorithms. In his general theory of inductive inference, Solomonoff uses the prior obtained by this formula, in Bayes' rule for prediction.

New Machine Learning Algorithms offer Safety and Fairness guarantees.

Statistics (math) - Algorithms, Direct Coding or Both?

Computer Algebra is a scientific area that refers to the study and development of algorithms and software for manipulating mathematical expressions and other mathematical objects. Although, properly speaking, computer algebra should be a subfield of scientific computing, they are generally considered as distinct fields because scientific computing is usually based on numerical computation with approximate floating point numbers, while symbolic computation emphasizes exact computation with expressions containing variables that have no given value and are manipulated as symbols. Software applications that perform symbolic calculations are called computer algebra systems, with the term system alluding to the complexity of the main applications that include, at least, a method to represent mathematical data in a computer, a user programming language (usually different from the language used for the implementation), a dedicated memory manager, a user interface for the input/output of mathematical expressions, a large set of routines to perform usual operations, like simplification of expressions, differentiation using chain rule, polynomial factorization, indefinite integration, etc. Computer algebra is widely used to experiment in mathematics and to design the formulas that are used in numerical programs. It is also used for complete scientific computations, when purely numerical methods fail, as in public key cryptography or for some non-linear problems.


What If


What If is introducing a conditional clause or on the condition or supposition that something will happen. In the event that a possibility will happen.

Suppose is to expect something to be true. To believe in something, especially on uncertain or tentative grounds. Take for granted or as a given and expect beforehand. Supposition is a message expressing an opinion based on incomplete evidence. A hypothesis that is taken for granted.

Stipulation is an assumption on which rests the validity or effect of something else.

Condition is an assumption on which rests the validity or effect of something else. Information that should be kept in mind when making a decision. A statement of what is required as part of an agreement. Condition is also a state at a particular time.

Precondition is an assumption on which rests the validity or effect of something else. An assumption that is taken for granted.

Probability (odds) - Variables - Decision Table

For Loop is a control flow statement for specifying iteration, which allows code to be executed repeatedly.

Artificial intelligence needs the "if" function, just like us. There are a lot of if's, with some if's that refer to other if's for more processing. Humorous What Ifs.

PHP or Hypertext Preprocessor is a general-purpose programming language originally designed for web development. PHP code may be executed with a command line interface (CLI), embedded into HTML code, or used in combination with various web template systems, web content management systems, and web frameworks. PHP code is usually processed by a PHP interpreter implemented as a module in a web server or as a Common Gateway Interface (CGI) executable. The web server combines the results of the interpreted and executed PHP code, which may be any type of data, including images, with the generated web page. PHP can be used for many programming tasks outside of the web context, such as standalone graphical applications and robotic drone control.

PHP if else - HP Control Structures

PHP - If Function (excel) - If Statement (excel)

PHP has the following conditional statements: if statement: executes some code only if a specified condition is True. if...else statement: executes some code if a condition is true and another code if the condition is False. if...elseif....else statement: selects one of several blocks of code to be executed. Switch statement: selects one of many blocks of code to be executed.


Logic Gates


Logic Gate is an idealized or physical device implementing a Boolean function; that is, it performs a logical operation on one or more binary inputs and produces a single binary output. Depending on the context, the term may refer to an ideal logic gate, one that has for instance zero rise time and unlimited fan-out, or it may refer to a non-ideal physical device. (see Ideal and real op-amps for comparison). Logic gates are primarily implemented using diodes or transistors acting as electronic switches, but can also be constructed using vacuum tubes, electromagnetic relays (relay logic), fluidic logic, pneumatic logic, optics, molecules, or even mechanical elements. With amplification, logic gates can be cascaded in the same way that Boolean functions can be composed, allowing the construction of a physical model of all of Boolean logic, and therefore, all of the algorithms and mathematics that can be described with Boolean logic. Logic circuits include such devices as multiplexers, registers, arithmetic logic units (ALUs), and computer memory, all the way up through complete microprocessors, which may contain more than 100 million gates. In modern practice, most gates are made from field-effect transistors (FETs), particularly metal–oxide–semiconductor field-effect transistorss (MOSFETs). Compound logic gates AND-OR-Invert (AOI) and OR-AND-Invert (OAI) are often employed in circuit design because their construction using MOSFETs is simpler and more efficient than the sum of the individual gates. In reversible logic, Toffoli gates are used. Neurons

Logic Gates (AND, OR, NOT) Basic Bolean
AND - IF (A=1 and B=1) Then OUT=1
OR - IF (A=1 or B=1) Then OUT=1
NOT - IF (A=1) Then OUT=0

AND gate is a basic digital logic gate that implements logical conjunction - it behaves according to the truth table to the right. A HIGH output (1) results only if all the inputs to the AND gate are HIGH (1). If none or not all inputs to the AND gate are HIGH, a LOW output results. The function can be extended to any number of inputs.

NAND gate is a logic gate which produces an output which is false only if all its inputs are true; thus its output is complement to that of the AND gate. A LOW (0) output results only if both the inputs to the gate are HIGH (1); if one or both inputs are LOW (0), a HIGH (1) output results. It is made using transistors and junction diodes. By De Morgan's theorem, AB=A+B, and thus a NAND gate is equivalent to inverters followed by an OR gate.

OR gate is a digital logic gate that implements logical disjunction – it behaves according to the truth table to the right. A HIGH output (1) results if one or both the inputs to the gate are HIGH (1). If neither input is high, a LOW output (0) results. In another sense, the function of OR effectively finds the maximum between two binary digits, just as the complementary AND function finds the minimum.

XOR gate is a digital logic gate that gives a true (1/HIGH) output when the number of true inputs is odd. An XOR gate implements an exclusive or; that is, a true output results if one, and only one, of the inputs to the gate is true. If both inputs are false (0/LOW) or both are true, a false output results. XOR represents the inequality function, i.e., the output is true if the inputs are not alike otherwise the output is false. A way to remember XOR is "one or the other but not both".

Inverter logic gate or NOT Gate is a logic gate which implements logical negation.

Subroutine (routines)
Code (computer programming)
Batch File (goals)
Binary (zeros and ones)
Iteration (developing ideas)
Software Design (computers)
Internet (combined intelligence)
Networks
Robots (building)

Conjunction (“and”)  -  Disjunction (“or”)  Exclusive Or  -  Negation (“not”) - Induction (deduction)

Gottfried Wilhelm Leibniz was a German polymath and philosopher (1716) who occupies a prominent place in the history of mathematics and the history of philosophy, having developed differential and integral calculus independently of Isaac Newton.

Characteristica Universalis is a universal and formal language imagined to express mathematical, scientific, and metaphysical concepts. Leibniz thus hoped to create a language usable within the framework of a universal logical calculation or calculus ratiocinator.

Calculus Ratiocinator is a theoretical universal logical calculation framework, a concept described in the writings of Gottfried Leibniz, usually paired with his more frequently mentioned characteristica universalis, a universal conceptual language.

Modulo Operation finds the remainder after division of one number by another (sometimes called modulus).

Modular Arithmetic is a system of arithmetic for integers, where numbers "wrap around" upon reaching a certain value—the modulus (plural moduli).

Mathematical Biophysics is a subfield of both biophysics and mathematical biology focusing of physical and physico-chemical mechanisms involved in physiological functions of living organisms, as well as the molecular structures supporting such physiological functions.

Our greatest intelligence now is already being formed by the Internet, which in some ways simulates the neural network of the human brain. But bringing together all our knowledge and information is only the beginning, because it will take the collective consensus of all the human brains in order for us to achieve intelligent solutions to our problems. And of course, incase of a major catastrophe, we will have to Secure our intelligence in something like the Global Seed Vault Because we would not want to start all over again as many humans civilizations had to do throughout human history. Backup our most important knowledge and information by transmitting it into space, store it in a satellite, store it on the moon and in multiple places. This we have to do. That's Intelligence.

NP-Hardness in computational complexity theory, is the defining property of a class of problems that are, informally, "at least as hard as the hardest problems in NP". A simple example of an NP-hard problem is the subset sum problem. A more precise specification is: a problem H is NP-hard when every problem L in NP can be reduced in polynomial time to H; that is, assuming a solution for H takes 1 unit time, we can use H‎'s solution to solve L in polynomial time. As a consequence, finding a polynomial algorithm to solve any NP-hard problem would give polynomial algorithms for all the problems in NP, which is unlikely as many of them are considered difficult. A common misconception is that the NP in "NP-hard" stands for "non-polynomial" when in fact it stands for "non-deterministic polynomial acceptable problems". Although it is suspected that there are no polynomial-time algorithms for NP-hard problems, this has not been proven. Moreover, the class P, in which all problems can be solved in polynomial time, is contained in the NP class.

NP is a complexity class used to classify decision problems. NP is the set of decision problems for which the problem instances, where the answer is "yes", have proofs verifiable in polynomial time.

Computational Complexity Theory

Decision Problem is a problem that can be posed as a yes-no question of the input values. An example of a decision problem is deciding whether a given natural number is prime. Another is the problem "given two numbers x and y, does x evenly divide y?". The answer is either 'yes' or 'no' depending upon the values of x and y. A method for solving a decision problem, given in the form of an algorithm, is called a decision procedure for that problem. A decision procedure for the decision problem "given two numbers x and y, does x evenly divide y?" would give the steps for determining whether x evenly divides y. One such algorithm is long division. If the remainder is zero the answer is 'yes', otherwise it is 'no'. A decision problem which can be solved by an algorithm is called decidable. Decision problems typically appear in mathematical questions of decidability, that is, the question of the existence of an effective method to determine the existence of some object or its membership in a set; some of the most important problems in mathematics are undecidable. The field of computational complexity categorizes decidable decision problems by how difficult they are to solve. "Difficult", in this sense, is described in terms of the computational resources needed by the most efficient algorithm for a certain problem. The field of recursion theory, meanwhile, categorizes undecidable decision problems by Turing degree, which is a measure of the noncomputability inherent in any solution. Decision Table.

Branch in computer science is an instruction in a computer program that can cause a computer to begin executing a different instruction sequence and thus deviate from its default behavior of executing instructions in order.[a] Branch (or branching, branched) may also refer to the act of switching execution to a different instruction sequence as a result of executing a branch instruction. Branch instructions are used to implement control flow in program loops and conditionals (i.e., executing a particular sequence of instructions only if certain conditions are satisfied). A branch instruction can be either an unconditional branch, which always results in branching, or a conditional branch, which may or may not cause branching depending on some condition. Also, depending on how it specifies the address of the new instruction sequence (the "target" address), a branch instruction is generally classified as direct, indirect or relative, meaning that the instruction contains the target address, or it specifies where the target address is to be found (e.g., a register or memory location), or it specifies the difference between the current and target addresses. Decision Tree.


Variables


Variable is something that is not consistent or having a fixed pattern. A value that is either arbitrary or not fully specified or unknown. Something that is subject to change which can make it difficult to predict.

Variation is an instance of change or the rate or magnitude of change. An activity that varies from a norm or standard.

Version is something a little different from others of the same type.

Variant is something a little different from others of the same type. Exhibiting variation and change. A variant in biology is a group of organisms within a species that differ in trivial ways from similar groups. A form or version of something that differs in some respect from other forms of the same thing or from a standard. A variable quantity that is random. An event that departs from expectations.

Variance is an event that departs from expectations. An activity that varies from a norm or standard. Variance in statistics is the second moment around the mean. The expected value of the square of the deviations of a random variable from its mean value. The quality or degree of being subject to variation. A difference between conflicting facts, claims or opinions. Discord that splits a group.

Mutation - Genetic Variations - Iteration - Unpredictable Phenomenon

Deviation is a variation that deviates from the standard or norm. The error of a compass due to local magnetic disturbances. Deviation in statistics is the difference between an observed value and the expected value of a variable or function.

Scenarios - Combination of Circumstances - Relative - Algorithm

Configurations is an arrangement of elements in a particular form, figure, or combination. Configurations in Chemistry is the fixed three-dimensional relationship of the atoms in a molecule, defined by the bonds between them. Configurations in Computing is the arrangement or set-up of the hardware and software that make up a computer system.

Design for X is when the X is a variable which can have one of many possible values. Safety Engineering.

Latent Variable are variables that are not directly observed but are rather inferred (through a mathematical model) from other variables that are observed (directly measured). Mathematical models that aim to explain observed variables in terms of latent variables are called latent variable models. Latent variable models are used in many disciplines, including psychology, economics, engineering, medicine, physics, machine learning/artificial intelligence, bioinformatics, natural language processing, econometrics, management and the social sciences. Sometimes latent variables correspond to aspects of physical reality, which could in principle be measured, but may not be for practical reasons. In this situation, the term hidden variables is commonly used (reflecting the fact that the variables are "really there", but hidden). Other times, latent variables correspond to abstract concepts, like categories, behavioral or mental states, or data structures. The terms hypothetical variables or hypothetical constructs may be used in these situations. One advantage of using latent variables is that they can serve to reduce the dimensionality of data. A large number of observable variables can be aggregated in a model to represent an underlying concept, making it easier to understand the data. In this sense, they serve a function similar to that of scientific theories. At the same time, latent variables link observable ("sub-symbolic") data in the real world to symbolic data in the modeled world.

Stochastic event or system is one that is unpredictable due to the influence of a random variable. The word "stochastic" comes from the Greek word στόχος (stokhos, "aim"). It occurs in a wide variety of professional and academic fields.

Random Variable in probability and statistics, a random variable, random quantity, aleatory variable or stochastic variable is a variable whose value is subject to variations due to chance (i.e. randomness, in a mathematical sense). A random variable can take on a set of possible different values (similarly to other mathematical variables), each with an associated probability, in contrast to other Mathematical Variables. Random - Hidden Variables.

Deterministic System is a system in which no randomness is involved in the development of future states of the system. A deterministic model will thus always produce the same output from a given starting condition or initial state.

Internalism and externalism are two opposing ways of explaining various subjects in several areas of philosophy. These include human motivation, knowledge, justification, meaning, and truth. The distinction arises in many areas of debate with similar but distinct meanings. Usually 'internalism' refers to the belief that an explanation can be given of the given subject by pointing to things which are internal to the person or their mind which is considering them. Conversely, externalism holds that it is things about the world which motivate us, justify our beliefs, determine meaning, etc.

Psychophysical is sharing the physical and psychological qualities.

Linearization refers to finding the linear approximation to a function at a given point.

Lyapunov optimization refers to the use of a Lyapunov function to optimally control a dynamical system. Lyapunov functions are used extensively in control theory to ensure different forms of system stability. The state of a system at a particular time is often described by a multi-dimensional vector. A Lyapunov function is a nonnegative scalar measure of this multi-dimensional state. Typically, the function is defined to grow large when the system moves towards undesirable states. System stability is achieved by taking control actions that make the Lyapunov function drift in the negative direction towards zero.

Variable and Attribute in research is a characteristic of an object (person, thing, etc.). Attributes are closely related to variables. A variable is a logical set of attributes. Variables can "vary" - for example, be high or low. How high, or how low, is determined by the value of the attribute (and in fact, an attribute could be just the word "low" or "high").

Variable in mathematics is an alphabetic character representing a number, called the value of the variable, which is either arbitrary or not fully specified or unknown. Making algebraic computations with variables as if they were explicit numbers allows one to solve a range of problems in a single computation. A typical example is the quadratic formula, which allows one to solve every quadratic equation by simply substituting the numeric values of the coefficients of the given equation to the variables that represent them.

Differentials (math)

Derivative of a function of a real variable measures the sensitivity to change of a quantity (a function value or dependent variable) which is determined by another quantity (the independent variable). Derivatives are a fundamental tool of calculus. For example, the derivative of the position of a moving object with respect to time is the object's velocity: this measures how quickly the position of the object changes when time is advanced.

Variable in computer science is a storage location paired with an associated symbolic name (an identifier), which contains some known or unknown quantity of information referred to as a value. The variable name is the usual way to reference the stored value; this separation of name and content allows the name to be used independently of the exact information it represents. The identifier in computer source code can be bound to a value during run time, and the value of the variable may thus change during the course of program execution.

Variable and Attribute in research is a characteristic of an object (person, thing, etc.). Attributes are closely related to variables. A variable is a logical set of attributes. Variables can "vary" - for example, be high or low. How high, or how low, is determined by the value of the attribute (and in fact, an attribute could be just the word "low" or "high".

Logistic Map is a polynomial mapping (equivalently, recurrence relation) of degree 2, often cited as an archetypal example of how complex, chaotic behaviour can arise from very simple non-linear dynamical equations.

Dynamical System is a system in which a function describes the time dependence of a point in a geometrical space. Examples include the mathematical models that describe the swinging of a clock pendulum, the flow of water in a pipe, and the number of fish each springtime in a lake.

Dependent and independent Variables dependent variables represent the output or outcome whose variation is being studied. The independent variables represent inputs or causes, i.e. potential reasons for variation.

Regression Analysis is a statistical process for estimating the relationships among variables. Patterns.

Symbolic Regression is a type of regression analysis that searches the space of mathematical expressions to find the model that best fits a given dataset, both in terms of accuracy and simplicity. No particular model is provided as a starting point to the algorithm. Instead, initial expressions are formed by randomly combining mathematical building blocks such as mathematical operators, analytic functions, constants, and state variables. (Usually, a subset of these primitives will be specified by the person operating it, but that's not a requirement of the technique.) Typically, new equations are then formed by recombining previous equations using genetic programming. By not requiring a specific model to be specified, symbolic regression isn't affected by human bias, or unknown gaps in domain knowledge. It attempts to uncover the intrinsic relationships of the dataset, by letting the patterns in the data itself reveal the appropriate models, rather than imposing a model structure that is deemed mathematically tractable from a human perspective. The fitness function that drives the evolution of the models takes into account not only error metrics (to ensure the models accurately predict the data), but also special complexity measures, thus ensuring that the resulting models reveal the data's underlying structure in a way that's understandable from a human perspective. This facilitates reasoning and favors the odds of getting insights about the data-generating system.

Variational Principle is a scientific principle used within the calculus of variations, which develops general methods for finding functions which extremize the value of quantities that depend upon those functions. For example, to answer this question: "What is the shape of a chain suspended at both ends?" we can use the variational principle that the shape must minimize the gravitational potential energy.

Condition Variable are synchronization primitives that enable threads to wait until a particular condition occurs. Condition variables are user-mode objects that cannot be shared across processes. Condition variables enable threads to atomically release a lock and enter the sleeping state.

Transient State is when a process variable or variables have been changed and the system has not yet reached a steady state. The time taken for the circuit to change from one steady state to another steady state is called the transient time. Transient analysis KVL and KCL to circuits containing energy storage elements results in differential. A transient process is a process in which process variables change over time. Transient analysis L and study to circuits containing energy storage elements results in differential. study of transient processes. Baseline.



Scenarios


Scenario is one of many known sequence of possible events. Real Life Examples.

Prepared for Emergencies - Planning - Cause and Effect - Relevance (situation)

Worst-Case Scenario is a concept in risk management wherein the planner, in planning for potential disasters, considers the most severe possible outcome that can reasonably be projected to occur in a given situation. Conceiving of worst-case scenarios is a common form of strategic planning, specifically scenario planning, to prepare for and minimize contingencies that could result in accidents, quality problems, or other issues. Variables - Patterns.

Minimax is a decision rule used in artificial intelligence, decision theory, game theory, statistics and philosophy for minimizing the possible loss for a worst case (maximum loss) scenario. When dealing with gains, it is referred to as "maximin"—to maximize the minimum gain. Originally formulated for two-player zero-sum game theory, covering both the cases where players take alternate moves and those where they make simultaneous moves, it has also been extended to more complex games and to general decision-making in the presence of uncertainty.

Conditional Probability is a measure of the probability of an event (some particular situation occurring) given that (by assumption, presumption, assertion or evidence) another event has occurred. If the event of interest is A and the event B is known or assumed to have occurred, "the conditional probability of A given B", or "the probability of A under the condition B", is usually written as P(A|B), or sometimes PB(A) or P(A/B). For example, the probability that any given person has a cough on any given day may be only 5%. But if we know or assume that the person has a cold, then they are much more likely to be coughing. The conditional probability of coughing given that you have a cold might be a much higher 75%.

Interactive Object-Based Media can help people understand different scenarios by giving people different options and other choices to choose from so they can see which decision was better.

How many questions deep do you need to go? How many levels? You can't prepare for everything, so how do you decide?

Formulating - Safety Engineering - Emerging Technologies

Exception Handling is the process of responding to the occurrence, during computation, of exceptions – anomalous or exceptional conditions requiring special processing – often changing the normal flow of program execution. It is provided by specialized programming language constructs or computer hardware mechanisms. Statistics.

Event Chain Methodology is an uncertainty modeling and schedule network analysis technique that is focused on identifying and managing events and event chains that affect project schedules. Event chain methodology is the next advance beyond critical path method and critical chain project management. Event chain methodology helps to mitigate the effect of motivational and cognitive biases in estimating and scheduling.

Preference Based Planning is a form of automated planning and scheduling which focuses on producing plans that additionally satisfy as many user-specified preferences as possible. In many problem domains, a task can be accomplished by various sequences of actions (also known as plans). These plans can vary in quality: there can be many ways to solve a problem but one generally prefers a way that is, e.g., cost-effective, quick and safe.

Regression Analysis is a statistical process for estimating the relationships among variables. It includes many techniques for modeling and analyzing several variables, when the focus is on the relationship between a dependent variable and one or more independent variables (or 'predictors'). More specifically, regression analysis helps one understand how the typical value of the dependent variable (or 'criterion variable') changes when any one of the independent variables is varied, while the other independent variables are held fixed.

Reasoning - Problem Solving - Quality Control

Scenario Planning is a strategic planning method that some organizations use to make flexible long-term plans. Part adaptation and generalization of classic methods.



Patterns


Pattern are elements that repeat in a predictable manner. A perceptual structure that can be observed and measured. Pattern is a discernible regularity in the world or in a manmade design that repeat in a predictable manner. A geometric pattern is a kind of pattern formed of geometric shapes and typically repeating like a wallpaper. Any of the senses may directly observe patterns. Conversely, abstract patterns in science, mathematics, or language may be observable only by analysis. Direct observation in practice means seeing visual patterns, which are widespread in nature and in art. Visual patterns in nature are often chaotic, never exactly repeating, and often involve fractals. Natural patterns include spirals, meanders, waves, foams, tilings, cracks, and those created by symmetries of rotation and reflection. Patterns have an underlying mathematical structure; indeed, mathematics can be seen as the search for regularities, and the output of any function is a mathematical pattern. Similarly in the sciences, theories explain and predict regularities in the world. In art and architecture, decorations or visual motifs may be combined and repeated to form patterns designed to have a chosen effect on the viewer. In computer science, a software design pattern is a known solution to a class of problems in programming. In fashion, the pattern is a template used to create any number of similar garments.

Pattern Recognition is the process of recognizing patterns by using machine learning algorithm or the automated recognition of patterns and regularities in data. Pattern recognition can be defined as the classification of data based on knowledge already gained or on statistical information extracted from patterns and/or their representation. IQ Test.

Time Based Patterns - Cycles - Sequences - Signals

Symmetry - Rhythm - Variables - Reference Pattern - Deciphering Code - Spatial Intelligence

Trend is a general direction in which something tends to move. 

Trend Estimation is a statistical technique to aid interpretation of data. When a series of measurements of a process are treated as a time series, trend estimation can be used to make and justify statements about tendencies in the data, by relating the measurements to the times at which they occurred.

Trend Analysis is the widespread practice of collecting information and attempting to spot a pattern. In project management, trend analysis is a mathematical technique that uses historical results to predict future outcome. In statistics, trend analysis often refers to techniques for extracting an underlying pattern of behavior in a time series which would otherwise be partly or nearly completely hidden by noise. If the trend can be assumed to be linear, trend analysis can be undertaken within a formal regression analysis, as described in Trend estimation. If the trends have other shapes than linear, trend testing can be done by non-parametric methods, e.g. Mann-Kendall test, which is a version of Kendall rank correlation coefficient. For testing and visualization of nonlinear trends also smoothing can be used.

Arithmetic Progression is a sequence of numbers such that the difference between the consecutive terms is constant.

Pattern Matching the act of checking a given sequence of tokens for the presence of the constituents of some pattern. In contrast to pattern recognition, the match usually has to be exact: "either it will or will not be a match." The patterns generally have the form of either sequences or tree structures. Uses of pattern matching include outputting the locations (if any) of a pattern within a token sequence, to output some component of the matched pattern, and to substitute the matching pattern with some other token sequence (i.e., search and replace). Sequence patterns (e.g., a text string) are often described using regular expressions and matched using techniques such as backtracking. Tree patterns are used in some programming languages as a general tool to process data based on its structure, e.g., Haskell, ML, Scala and the symbolic mathematics language Mathematica have special syntax for expressing tree patterns and a language construct for conditional execution and value retrieval based on it. For simplicity and efficiency reasons, these tree patterns lack some features that are available in regular expressions. Often it is possible to give alternative patterns that are tried one by one, which yields a powerful conditional programming construct. Pattern matching sometimes includes support for guards. Term rewriting and graph rewriting languages rely on pattern matching for the fundamental way a program evaluates into a result.

Pattern Separation is defined as the process by which overlapping or similar inputs (representations) are transformed into less similar outputs whereas pattern completion is the reconstruction of complete stored representations from partial inputs that are part of the stored representation.

Computers can read zero's and ones, which means they can be taught to look for patterns. And when these patterns are labeled correctly and accurately, a computer can identify things in the world pretty much the same way as humans do.

Software Design Pattern is a reusable solution to a commonly occurring problem within a given context in software design. A template for how to solve a problem that can be used in many different situations. Object-oriented design patterns typically show relationships and interactions between classes or objects, without specifying the final application classes or objects that are involved.

Command Pattern is a behavioral design pattern in which an object is used to encapsulate all information needed to perform an action or trigger an event at a later time. This information includes the method name, the object that owns the method and values for the method parameters. Grid.

Deep learning is great for finding Trends and Patterns in Data. But if you don't use this information to benefit society, then we will continue to suffer, as we are now.

Sensors - Time

"Computers will help us make better Predictions, Ai will also help us make better Decisions, but Humans still have to steer."

Data Dredging is the use of data mining to uncover patterns in data that can be presented as statistically significant, without first devising a specific hypothesis as to the underlying causality. (also known as data fishing, data snooping, and p-hacking).

Linear Discriminant Analysis is a generalization of Fisher's linear discriminant, a method used in statistics, pattern recognition and machine learning to find a linear combination of features that characterizes or separates two or more classes of objects or events. The resulting combination may be used as a linear classifier, or, more commonly, for dimensionality reduction before later classification.

Facial Recognition System is a computer application capable of identifying or verifying a person from a digital image or a video frame from a video source. One of the ways to do this is by comparing selected facial features from the image and a face database. Body Language.

Composite Entity Pattern is used to model, represent, and manage a set of interrelated persistent objects rather than representing them as individual fine-grained entity beans, and also a composite entity bean represents a graph of objects.

Statistics (math)

Cycle Detection is the algorithmic problem of finding a cycle in a sequence of iterated function values.

Vibrations (hz)

Trypophobia is the irrational fear of irregular patterns or clusters of small holes or bumps.

Profiling algorithms or mathematical techniques allow the discovery of patterns or correlations in large quantities of data.

What happens in the sensory cortex when learning and recognizing patterns. Learning to recognize a specific pattern or sequence, and process the ensuing actions and outcomes, involves neurons across the whole cortex. An animal does not, it seems, sense the world separately from what it needs to feel in order to guide behavior.

How our brains know when something's different. Scientists discovered how a set of high frequency brain waves may help us unconsciously know when something's different by comparing memories of the past with present experiences. Almost every experience we store into memory can be used to set our expectations and predictions for the future. Predictive coding basically states that the brain optimizes neural activity for processing information. The brain uses more neural activity to process new information than it does for things that we are familiar with. It takes just one moment to not only remember a new experience but also to use memories of that experience to set future expectations.

Predictive Coding is a theory of brain function in which the brain is constantly generating and updating a mental model of the environment. The model is used to generate predictions of sensory input that are compared to actual sensory input. This comparison results in prediction errors that are then used to update and revise the mental model.

New research describes a new model for how the brain interprets patterns in complex networks. They found that the ability to detect patterns stems in part from the brain's desire to represent things in the simplest way possible and that the brain is constantly weighing the pressures of complexity and simplicity when making decisions. The human brain is a highly advanced information processor composed of more than 86 billion neurons. Humans are adept at recognizing patterns from complex networks, such as languages, without any formal instruction. Previously, cognitive scientists tried to explain this ability by depicting the brain as a highly optimized computer, but there is now discussion among neuroscientists that this model might not accurately reflect how the brain works.



Sensors


Sensor is an object whose purpose is to detect events or detect changes or signals in its environment and then sends the information to the computer which then tells the actuator or the output devices to provide the corresponding output. A sensor is a device that converts real world data or analog into data that a computer can understand using ADC or analog to digital converter. All living organisms contain biological sensors with functions similar to those of the mechanical devices described. Most of these are specialized cells that are sensitive to: Light, motion, temperature, magnetic fields, gravity, humidity, moisture, vibration, pressure, electrical fields, sound, and other physical aspects of the external environment. Physical aspects of the internal environment, such as stretch, motion of the organism, and position of appendages (proprioception). Estimation of biomolecules interaction and some kinetics parameters. Internal metabolic indicators, such as glucose level, oxygen level, or osmolality. Internal signal molecules, such as hormones, neurotransmitters, and cytokines. Differences between proteins of the organism itself and of the environment or alien creatures. DIY Science Kits - Machine Learning.

Sensor Grid integrates wireless sensor networks with grid computing concepts to enable real-time sensor data collection and the sharing of computational and storage resources for sensor data processing and management. It is an enabling technology for building large-scale infrastructures, integrating heterogeneous sensor, data and computational resources deployed over a wide area, to undertake complicated surveillance tasks such as environmental monitoring.

Polymetric involves using more than one meter. - Sensor Array - Touch Bionics.

Sensor Fusion is combining of sensory data or data derived from disparate sources such that the resulting information has less uncertainty than would be possible when these sources were used individually. The term uncertainty reduction in this case can mean more accurate, more complete, or more dependable, or refer to the result of an emerging view, such as stereoscopic vision (calculation of depth information by combining two-dimensional images from two cameras at slightly different viewpoints). The data sources for a fusion process are not specified to originate from identical sensors. One can distinguish direct fusion, indirect fusion and fusion of the outputs of the former two. Direct fusion is the fusion of sensor data from a set of heterogeneous or homogeneous sensors, soft sensors, and history values of sensor data, while indirect fusion uses information sources like a priori knowledge about the environment and human input. Sensor fusion is also known as (multi-sensor). Data Fusion is the process of integrating multiple data sources to produce more consistent, accurate, and useful information than that provided by any individual data source. Information Fusion is the merging of information from heterogeneous sources with differing conceptual, contextual and typographical representations. Filtering - Cross Talk.

Wireless Sensor Network are spatially distributed autonomous sensors to monitor physical or environmental conditions, such as temperature, sound, pressure, etc. and to cooperatively pass their data through the network to other locations.

Sensor Web is a type of sensor network that is especially well suited for environmental monitoring.

Next generation of Networked Smart Devices can communicate directly with one another without human intervention. It needs only a very small amount of power to maintain this constant listening and always be on the alert, so it still saves energy overall while extending the battery life of the larger device. A well-designed wake-up receiver also allows the device to be turned on from a significant distance. A sleeping device can still suck the life out of a battery. A wake-up receiver that turns on a device in response to incoming ultrasonic signals -- signals outside the range that humans can hear. By working at a significantly smaller wavelength and switching from radio waves to ultrasound, this receiver is much smaller than similar wake-up receivers that respond to radio signals, while operating at extremely low power and with extended range.

Enabling 'Internet of Photonic Things' with Miniature Sensors. Swapping electrons for photons, researchers have developed wireless sensors which are not subject to electromagnetic interference and are smaller and generally more flexible than the currently electronics-based technology. Records environmental data using a wireless photonic sensor resonator with a whispering-gallery-mode (WGM) architecture. Whispering-Gallery Wave are a type of wave that can travel around a concave surface. they can exist for light and for other waves, with important applications in nondestructive testing, lasing, cooling and sensing, as well as in astronomy.

Batteryless smart devices closer to reality. Researchers have taken a huge step towards making smart devices that do not use batteries or require charging. An RFID tag is modified by cutting out a small part its antenna (silver ribbon) and placing a small light-sensing phototransistor or temperature-responsive resistor (thermistor) on it.

RFID stands for Radio Frequency Identification, which uses electromagnetic fields to automatically identify and track tags attached to objects. The tags contain electronically-stored information. Passive tags collect energy from a nearby RFID reader's interrogating radio waves. Telemetry - Bar Codes.

MIT Engineers configure RFID Tags to work as Sensors. Platform may enable continuous, low-cost, reliable devices that detect chemicals in the environment.

Sensor Fusion is combining of sensory data or data derived from disparate sources such that the resulting information has less uncertainty than would be possible when these sources were used individually.

Smart machine components alert users to damage and wear. Technology to create polymer-bonded magnets with intricate geometries and arbitrary shapes, opening up new possibilities for manufacturing and product design.

Quantum Sensors. Researchers have created a chip on which laser light interacts with a tiny cloud of atoms to serve as a miniature toolkit for measuring important quantities such as length with quantum precision. The design could be mass-produced with existing technology.

Quantum Sensor is a device that exploits quantum correlations, such as quantum entanglement, to achieve a sensitivity or resolution that is better than can be achieved using only classical systems. A quantum sensor can measure the effect of the quantum state of another system on itself. The mere act of measurement influences the quantum state and alters the probability and uncertainty associated with its state during measurement. The Defense Advanced Research Projects Agency has recently launched a research program in optical quantum sensors that seeks to exploit ideas from quantum metrology and quantum imaging, such as quantum lithography and the NOON state, in order to achieve these goals with optical sensor systems such as Lidar. Quantum sensor is also a term used in other settings where entangled quantum systems are exploited to make better atomic clocks or more sensitive magnetometers. A good example of an early quantum sensor is an APD avalanche photodiode such as the AD500-8 TO52S1 as these have been used to detect entangled photons and in fact with additional cooling and sensor improvements can be used where PMTs once ruled the market such as medical imaging. These are also now being used by First Sensor in the form of 2-D and even 3-D stacked arrays as a direct replacement for conventional sensors based on silicon diodes.

Wearable Sensors - Flexible Electronics - Bio-Monitoring - Health Monitors - Smart Homes

Scientists Invent Threads to Detect Gases when Woven into Clothing. A novel fabrication method to create dyed threads that change color when they detect a variety of gases. Woven into clothing, smart, gas-detecting threads could provide a reusable, washable, and affordable safety asset in medical, workplace, military and rescue environments. The study describes the fabrication method and its ability to extend to a wide range of dyes and detection of complex gas mixtures.

Electric Cloth: Flexible, wearable supercapacitors based on porous nanocarbon nanocomposites.

Tissue Paper Sensors show promise for health care, entertainment, Robotics - Bionics

Wearable Ultrasound Patch Monitors Blood Pressure Deep Inside Body that non-invasively monitors blood pressure in arteries deep beneath the skin that could help people detect cardiovascular problems earlier on and with greater precision. In tests, the patch performed as well as some clinical methods to measure blood pressure.

Simple Stickers that are Wearable Medical Electronic Devices that someone can easily attach to their skin that could save lives of patients, athletes and lower medical costs. The devices are made out of paper to lower the cost of personalized medicine. Telemedicine (doctor house calls over the internet).

Stanford engineers have developed wireless sensors that stick to the skin to track our health. Stanford engineers have developed experimental stickers that pick up physiological signals emanating from the skin, then wirelessly beam these health readings to a receiver clipped onto clothing. It’s all part of a system called BodyNet.

Wearable sensors detect what's in your sweat. New easy-to-make sensors can provide real-time measurements of sweat rate and electrolytes and metabolites in perspiration.

Researchers have developed an implantable, biofuel-powered sensor that runs on sugar and can monitor a body's biological signals to detect, prevent and diagnose diseases.

Implantable biosensor operates without batteries. Researchers have revealed their new biodegradable motion sensor -- paving the way for implanted nanotechnology that could help future sports professionals better monitor their movements to aid rapid improvements, or help caregivers remotely monitor people living with dementia. The ATI's technology builds on its previous work around triboelectric nanogenerators (TENG), where researchers used the technology to harness human movements and generate small amounts of electrical energy. Combining the two means self-powered sensors are possible without the need for chemical or wired power sources.

A Self-Powered Heart Monitor taped to the Skin called an organic electrochemical transistor, is a human-friendly, ultra-flexible organic sensor powered by sunlight. A type of electronic device that can be used to measure a variety of biological functions.

Materials scientist creates fabric alternative to batteries for wearable devices. A method for making a charge-storing system that is easily integrated into clothing for 'embroidering a charge-storing pattern onto any garment.' uses a micro-supercapacitor and combines vapor-coated conductive threads with a polymer film, plus a special sewing technique to create a flexible mesh of aligned electrodes on a textile backing. The resulting solid-state device has a high ability to store charge for its size, and other characteristics that allow it to power wearable biosensors.

Sensors that are literally ‘music to one’s ears’. Researchers have found a new use for a 3,000-year-old African musical instrument: detecting toxic substances and counterfeit medications. The sensor, based on the mbira (pronounced 'em-bir'-uh') is inexpensive and easy to operate, allowing its use in developing regions, the researchers say.

Scientists develop new tool for imprinting biochips. New technology could allow researchers to fit more biochemical probes onto a single biochip and reduce the cost of screening and analyzing changes biochips (also known as microarrays), which are used to screen for and analyze biological changes associated with disease development, biothreat agents, pollution, toxins and other areas of research that involve biological components.

A nanotech sensor that turns molecular fingerprints into bar codes. A new system can detect and analyze molecules without the need for an infrared spectrometer. The system uses nanostructured metapixels to detect and then translate molecules' unique signatures into bar codes. The technology can be integrated into a compact sensor chip. It opens the door to large-scale image-based detection of materials using machine-learning technology.

Route to Flexible Electronics made from Exotic Materials. Cost-effective method produces semiconducting films from materials that outperform silicon. researchers fabricated flexible films made from gallium arsenide, gallium nitride, and lithium fluoride -- materials that exhibit better performance than silicon but until now have been prohibitively expensive to produce in functional devices. MIT researchers have devised a way to grow single crystal GaN thin film on a GaN substrate through two-dimensional materials. The GaN thin film is then exfoliated by a flexible substrate, showing the rainbow color that comes from thin film interference. This technology will pave the way to flexible electronics and the reuse of the wafers.

Molecular Probe is a group of atoms or molecules used in molecular biology or chemistry to study the properties of other molecules or structures. If some measurable property of the molecular probe used changes when it interacts with the analyte (such as a change in absorbance), the interactions between the probe and the analyte can be studied. This makes it possible to indirectly study the properties of compounds and structures which may be hard to study directly. The choice of molecular probe will depend on which compound or structure is being studied as well as on what property is of interest. Radioactive DNA or RNA sequences are used in molecular genetics to detect the presence of a complementary sequence by molecular hybridization.

Biochip are essentially miniaturized laboratories that can perform hundreds or thousands of simultaneous biochemical reactions. Biochips enable researchers to quickly screen large numbers of biological analytes for a variety of purposes, from disease diagnosis to detection of bioterrorism agents. Digital microfluidic biochips have become one of the most promising technologies in many biomedical fields. In a digital microfluidic biochip, a group of (adjacent) cells in the microfluidic array can be configured to work as storage, functional operations, as well as for transporting fluid droplets dynamically.

Plasmonic Nanoantenna Arrays could lead to the development of a new generation of ultrasensitive and low-cost fluorescence sensors that could be used to monitor water quality.

Ultra-sensitive sensor with gold nanoparticle array. Scientists have developed a new type of sensor platform using a gold nanoparticle array, which is 100 times more sensitive than current similar sensors for air pollution or for medical diagnostics. The sensor is made up of a series of gold disk-shaped nanoparticles on a glass slide.

UW team shatters long-range communication barrier for devices that consume almost no power. Sensor allows devices that run on extremely low power for the first time to communicate over long distances.

Force-Sensing Resistor is a material whose resistance changes when a force or pressure is applied.

New air-pressure sensor could improve everyday devices. Micro-electro-mechanical system - micro-switch.

Engineers Create Artificial Skin That "Feels" Temperature Changes.

Paving the way for sensor interfaces that are 30 times smaller. Researchers have invented a novel class of Digital-to-Analog (DAC) and Analog-to-Digital Converters (ADC) that can be entirely designed with a fully-automated digital design methodology.

Human Senses - Materials Science (strength limits)

"Sensors could make people vulnerable if people don't understand the function or the process of sensors. If people don't know how to sense the world without using artificial sensors, then people will be vulnerable to failure when the sensors are not available or not working correctly."

New Malleable 'Electronic Skin' Self-Healable and Recyclable. Electronic skin, known as e-skin, is a thin, translucent material that can mimic the function and mechanical properties of human skin that can measure pressure, temperature, humidity and air flow.

Artificial 'skin' gives robotic hand a sense of touch. UH researchers discover new form of stretchable electronics, sensors and skins. Bionics.

The sensitive strain sensor that can detect the weight of a feather.

Repetition key to Self-Healing, Flexible Medical Devices. Medical devices powered by synthetic proteins created from repeated sequences of proteins may be possible, according to materials science and biotechnology experts, who looked at material inspired by the proteins in squid ring teeth.

Biosensor is an analytical device, used for the detection of an analyte, that combines a biological component with a physicochemical detector. The sensitive biological element (e.g. tissue, microorganisms, organelles, cell receptors, enzymes, antibodies, nucleic acids, etc.) is a biologically derived material or biomimetic component that interacts (binds or recognizes) with the analyte under study. Miniature Technology, Big Hope for Disease Detection

Biosensors will be inexpensive, do more, go everywhere.

Food Sensors (hand held) - Sensors for Measuring Soil Moisture

Compact Fiber Optic Sensor offers sensitive analysis in narrow spaces. Compact sensor would be useful for biomedical, chemical and food safety applications. Researchers have developed a new flexible sensor with high sensitivity that is designed to perform variety of chemical and biological analyses in very small spaces.

Chemical Sensor is a self-contained analytical device that can provide information about the chemical composition of its environment, that is, a liquid or a gas phase. The information is provided in the form of a measurable physical signal that is correlated with the concentration of a certain chemical species (termed as analyte).

Chemical Sensors (PDF)

Nanosensor are any biological, chemical, or surgical sensory points used to convey information about nanoparticles to the macroscopic world.

Synthetic Sensors: Towards General-Purpose Sensing (youtube)

Synthesizing Chemical-Sensing Cells from scratch. Scientists create artificial cells that can express distinct genes in response to specific chemical signals. The scientists chose histamine as the chemical signal for their artificial cells because it is an important biological compound in the immune system. If you feel an itch, histamine is the likely culprit. It is also released by the body during allergic reactions and helps defend against foreign pathogens by spurring inflammation. To detect histamine, they created a molecule called an RNA aptamer. RNA aptamers are small segments of RNA building blocks that can be engineered to act as binding agents to specific target molecules. Next, the team developed a so-called riboswitch that would turn this signal detection into action -- specifically, translating a gene to produce a protein. Normally, cells produce proteins when templates made of messenger RNA (mRNA) bind to cellular structures called ribosomes. Here, the scientists used the histamine aptamer to design a riboswitch that alters the shape of the mRNA upon binding histamine. In the absence of histamine, the shape of the mRNA prevents the ribosome from binding, and no protein is produced. Histamine-bound mRNA, however, allows ribosome to bind and synthesize proteins.

Wearable, Low-Cost Sensor to Measure Skin Hydration

Metal Printing Offers Low-Cost Way to Make Flexible, Stretchable Electronics

How a $10 Microchip Turns 2-D Ultrasound Machines to 3-D Imaging Devices (youtube)

Chip-Based Sensors with incredible sensitivity used for motion, temperature, pressure or biochemical sensing. sensor consists of solid spheres.

Swallowable Sensors reveal mysteries of Human Gut Health.

Smart Homes - Smartphone Accessories - Medical Sensors

Sounds of action: Using ears, not just eyes, improves robot perception. People rarely use just one sense to understand the world, but robots usually only rely on vision and, increasingly, touch. Researchers find that robot perception could improve markedly by adding another sense: hearing.


Vision Sensors - Image Recognition - Computer Sensors that See


Computer Vision deals with how computers can be made to gain high-level understanding from digital images or videos. From the perspective of engineering, it seeks to automate tasks that the human visual system can do. Image Recognition.

IBM believes computers will be able to identify images and understand what they mean without the use of tags. This will lead to systems that can help doctors analyze X-ray pictures, magnetic resonance imaging (MRI) machine, ultrasound or computerized tomography scans. 

Activity Recognition (facial recognition) - Optical Character Recognition

How computers learn to recognize objects instantly (video and interactive text)
Build a TensorFlow Image Classifier in 5 Min (youtube)

Image Sensor is a sensor that detects and conveys the information that constitutes an image. It does so by converting the variable attenuation of light waves (as they pass through or reflect off objects) into signals, small bursts of current that convey the information. The waves can be light or other electromagnetic radiation. Image sensors are used in electronic imaging devices of both analog and digital types, which include digital cameras, camera modules, medical imaging equipment, night vision equipment such as thermal imaging devices, radar, sonar, and others. As technology changes, digital imaging tends to replace analog imaging. Charge-Coupled Device or CCD is a device for the movement of electrical charge, usually from within the device to an area where the charge can be manipulated, for example conversion into a digital value. This is achieved by "shifting" the signals between stages within the device one at a time. CCDs move charge between capacitive bins in the device, with the shift allowing for the transfer of charge between bins.

Visual Search Engine is a search engine designed to search for information on the World Wide Web through the input of an image or a search engine with a visual display of the search results. Information may consist of web pages, locations, other images and other types of documents. This type of search engines is mostly used to search on the mobile Internet through an image of an unknown object (unknown search query). Examples are buildings in a foreign city. These search engines often use techniques for Content Based Image Retrieval.

Imagenet is an image database organized according to the WordNet hierarchy (currently only the nouns), in which each node of the hierarchy is depicted by hundreds and thousands of images. Currently we have an average of over five hundred images per node.

Tensorflow Image Recognition QuocNet, AlexNet, Inception (GoogLeNet), BN-Inception-v2. Tensor Flow Open Source Software Library for Machine Intelligence. Machine Learning.

Vicarious is developing machine learning software based on the computational principles of the human brain. Known as the Recursive Cortical Network (RCN), it is a visual perception system that interprets the contents of photographs and videos in a manner similar to humans. Recaptcha (google) - CAPTCHA

Object Recognition technology in the field of computer vision for finding and identifying objects in an image or video sequence. Humans recognize a multitude of objects in images with little effort, despite the fact that the image of the objects may vary somewhat in different view points, in many different sizes and scales or even when they are translated or rotated. Objects can even be recognized when they are partially obstructed from view. This task is still a challenge for computer vision systems. Many approaches to the task have been implemented over multiple decades.

PASCAL - Pattern Analysis, Statistical Modelling and Computational Learning
VOC - Visual Object Classes

Machine Learning now can outperform dermatologists at recognizing skin cancers in blemish photos. They can beat cardiologists in detecting arrhythmias in EKGs.

Visual Search Engine App - Gif

Arxiv Full Resolution Image Compression with Recurrent Neural Networks.

QIS is the next generation of image sensors where high-speed single-photon detection is used to unlock new image capture capabilities for consumers and professionals not possible with today’s devices. "Jot" is the specialized pixel that is sensitive enough to detect a single photon of light. Revolutionary detection technologies are developed to enable accurate photon-counting at room temperature without the use of electron avalanche multiplication.

Orientation in computer vision and image processing is when a sufficiently small image regions can be characterized as locally one-dimensional, e.g., in terms of lines or edges. For natural images this assumption is usually correct except at specific points, e.g., corners or line junctions or crossings, or in regions of high frequency textures. However, what size the regions have to be in order to appear as one-dimensional varies both between images and within an image. Also, in practice a local region is never exactly one-dimensional but can be so to a sufficient degree of approximation.

Dynamic Projection Mapping onto Deforming non-rigid Surface at 1,000 fps with 3 ms delay (youtube)
High-Speed Projector DynaFlash

Researchers teach computers to see Optical Illusions. By making a neural-network computer model that can be fooled by optical illusions like humans, the researchers advanced knowledge of the human visual system and may help improve artificial vision.

Projection Mapping is a projection technology used to turn objects, often irregularly shaped, into a display surface for video projection. These objects may be complex industrial landscapes, such as buildings, small indoor objects or theatrical stages. By using specialized software, a two- or three-dimensional object is spatially mapped on the virtual program which mimics the real environment it is to be projected on. The software can interact with a projector to fit any desired image onto the surface of that object. This technique is used by artists and advertisers alike who can add extra dimensions, optical illusions, and notions of movement onto previously static objects. The video is commonly combined with, or triggered by, audio to create an audio-visual narrative.

Machine Vision is the technology and methods used to provide imaging-based automatic inspection and analysis for such applications as automatic inspection, process control, and robot guidance, usually in industry. Machine vision is a term encompassing a large number of technologies, software and hardware products, integrated systems, actions, methods and expertise. Machine vision as a systems engineering discipline can be considered distinct from computer vision, a form of computer science. It attempts to integrate existing technologies in new ways and apply them to solve real world problems. The term is also used in a broader sense by trade shows and trade groups; this broader definition also encompasses products and applications most often associated with image processing.

LIDAR-Sensing using Light. - Self Driving Cars - Navigation.

Acoustic Non-Line-of-Sight Imaging is a novel approach to seeing around corners using acoustic echoes. A system of speakers emits sound waves which scatter from a wall to a hidden object and back.

New Depth Sensors could make self-driving cars practical Computational method improves resolution of time-of-flight depth sensors 1,000-fold.

A marriage of Light-Manipulation Technologies, researchers have built a metasurface-based lens atop a Micro-Electro-Mechanical System (MEMS) platform. The result is a new, infrared light-focusing system that combines the best features of both technologies while reducing the size of the optical system. combining the strengths of high-speed dynamic control and precise spatial manipulation of wave fronts. (Metalenses).

Bioinspired camera could help self-driving cars see better by detecting a property of light known as polarization and featuring a dynamic range about 10,000 times higher than today's commercial cameras. Dynamic range is a measure of the brightest and darkest areas a camera can capture simultaneously. With these, the camera can see better in driving conditions such as the transition from a dark tunnel into bright sunlight or during hazy or foggy conditions.

Camera brings unseen world to light. Portable polarization-sensitive camera could be used in machine vision, autonomous vehicles, security, atmospheric chemistry and more. Researchers have developed a highly compact, portable camera that can image polarization in a single shot. The miniature camera -- about the size of a thumb -- could find a place in the vision systems of autonomous vehicles, onboard planes or satellites to study atmospheric chemistry, or be used to detect camouflaged objects.

Successful automatic landing with vision assisted navigation. Researchers have now demonstrated a completely automatic landing with vision assisted navigation that functions properly without the need for ground-based systems. C2Land.

Micro-Electro-Mechanical Systems is the technology of microscopic devices, particularly those with moving parts. It merges at the nano-scale into nanoelectromechanical systems (NEMS) and nanotechnology. MEMS are also referred to as micromachines in Japan, or micro systems technology (MST) in Europe.

Inception Model Image Recognition (tensorflow)

Statistical Classification is the problem of identifying to which of a set of categories (sub-populations) a new observation belongs, on the basis of a training set of data containing observations (or instances) whose category membership is known.

Computer Algorithm that is nearly as accurate as people are at Image Analysis of brain circuitry and neural networks.

Vrad (Radiology) - Zebra-Med (medical image diagnosis).

Minimalist Machine Learning Algorithms Analyze Images from Very Little Data. CAMERA researchers develop highly efficient neural networks for analyzing experimental scientific images from limited training data.

Convolutional Neural Network is a class of deep, feed-forward artificial neural network that have successfully been applied to analyzing visual imagery.

Transfer Learning or inductive transfer is a research problem in machine learning that focuses on storing knowledge gained while solving one problem and applying it to a different but related problem. For example, knowledge gained while learning to recognize cars could apply when trying to recognize trucks. This area of research bears some relation to the long history of psychological literature on transfer of learning, although formal ties between the two fields are limited.

Optical Character Recognition is the mechanical or electronic conversion of images of typed, handwritten or printed text into machine-encoded text, whether from a scanned document, a photo of a document, a scene-photo (for example the text on signs and billboards in a landscape photo) or from subtitle text superimposed on an image (for example from a television broadcast).

Image Classification Algorithm

Neural hardware for image recognition in nanoseconds. Usually, artificial intelligence is based on software. Scientists created intelligent hardware, which is much faster. Within nanoseconds, the chip can analyze images and provide the correct output.

Translations (language)

Question and Answer Platforms

NeuralTalk is a Python+numpy project for learning Multimodal Recurrent Neural Networks that describe images with sentences

Computers being able to identify images means that blind people will see by way of  Sensory Substitution.

David Eagleman: Can we Create new Senses for Humans (video)
Sensory Vest

Sight Tools (seeing)

Weakness in modern computer vision systems

Advances in Technology Provide Clearer Insight Into Brain's Visual System. A new high-density EEG can capture the brain's neural activity at a higher spatial resolution than ever before. This next generation brain-interface technology is the first non-invasive, high-resolution system of its kind, providing higher density and coverage than any existing system.

Metalens combined with an Artificial Muscle. Artificial eye automatically stretches to simultaneously focus and correct astigmatism and image shift. Metalens is a lens made from a metamaterial, which is any material that obtains its electromagnetic properties from its structure rather than from its chemical composition; especially a material engineered to have features of a size less than that of the wavelength of a class of electromagnetic radiation.

Ghost Imaging is a technique that produces an image of an object by combining information from two light detectors: a conventional, multi-pixel detector that doesn't view the object, and a single-pixel (bucket) detector that does view the object. Two techniques have been demonstrated. A quantum method uses a source of pairs of entangled photons, each pair shared between the two detectors, while a classical method uses a pair of correlated coherent beams without exploiting entanglement. Both approaches may be understood within the framework of a single theory. (also called "coincidence imaging", "two-photon imaging" or "correlated-photon imaging").

Researchers capture moving object with ghost imaging. Researchers have developed a way to capture moving objects with the unconventional imaging method known as ghost imaging. The new method could make the imaging technique practical for new applications such as biomedical imaging, security checks and video compression and storage. Ghost imaging comes with a host of advantages, one of which is that it allows one to form an image by illuminating the object with lower light levels than traditional imaging approaches. However, ghost imaging has been limited to stationary objects because it takes a long time to project the sequence of light patterns onto the object that is necessary to reconstruct an image. This causes images of a moving object to appear blurry.

Adversarial Reprogramming of Neural Networks.

When a machine can see the world in the same way that a human does, then we will have some really cool robots.

Light processing improves robotic sensing, study finds. A team of researchers uncovered how the human brain processes bright and contrasting light, which they say is a key to improving robotic sensing and enabling autonomous agents to team with humans.

Breakthrough optical sensor mimics human eye. Previous attempts to build a human-eye type of device, called a retinomorphic sensor, have relied on software or complex hardware. But the new sensor's operation is part of its fundamental design, using ultrathin layers of perovskite semiconductors -- widely studied in recent years for their solar energy potential -- that change from strong electrical insulators to strong conductors when placed in light. The new sensor could be a perfect match for the neuromorphic computers that will power the next generation of artificial intelligence in applications like self-driving cars, robotics and advanced image recognition, Labram said. Unlike traditional computers, which process information sequentially as a series of instructions, neuromorphic computers are designed to emulate the human brain's massively parallel networks.


Hearing Sensors - Computer Sensors that Listen


There will also be improvements in computers' ability to hear and understand sound. Greater sensitivity to sound pressure, vibrations and waves could lead to more-accurate landslide warnings, for example.

Speech Recognition methodologies and technologies that enables the recognition and translation of spoken language into text by computers. It is also known as "automatic speech recognition" (ASR), "computer speech recognition", or just "speech to text" (STT). It incorporates knowledge and research in the linguistics, computer science, and electrical engineering fields. Your voice is measured by Frequency, the Wavelengths of Sound at a specific moment. Software breaks down your command into 25-millisecond slivers, then converts each wavelength measurement into digestible numbers. The software compares those sonic signatures to its catalog of sounds until its confidence scores are high enough that it can assume that you said. The software compares the words it thinks you've said to its stores of example sentences, which inform how it understands syntax and vocabulary. Acoustic and language models constantly adjust to how people use them. That's where A.I., specifically machine learning, comes in context to be more accurate.

YouTube built an Automated Content Detection System that prevents most unauthorized clips from appearing on its site.

Artificial intelligence produces realistic sounds that fool humans Video-trained system from MIT’s Computer Science and Artificial Intelligence Lab could help robots understand how objects interact with the world. 

Machine-Learning System Processes Sounds like Humans do.

Cloud DX uses AI technology to scrutinize the audio waveform of a human cough, which allows it to detect asthma, tuberculosis, pneumonia, and other lung diseases.

Selecting sounds: How the brain knows what to listen to. New noninvasive approach reveals brain mechanisms of auditory attention.


Taste - Flavor Sensors


Computers with virtual taste buds will be able to Calculate flavor, according to IBM, helping chefs improve recipes or create new ones. The systems will break down ingredients to their respective chemicals and calculate their interactions with neural sensors in a person's tongue and nose.


Smell - Oder Sensors


According to IBM, computers will have an acute sense of smell in order to diagnose from a person's breath a coming cold, liver and kidney disorders, diabetes and tuberculosis. Similar to how a Breathalyzer detects alcohol, the computer will be able to check for molecular biomarkers pointing to diseases. 

Machine Olfaction is the automated simulation of the sense of smell.

Electronic Nose is a device intended to detect odors or flavors. Over the last decade, "electronic sensing" or "e-sensing" technologies have undergone important developments from a technical and commercial point of view. The expression "electronic sensing" refers to the capability of reproducing human senses using sensor arrays and pattern recognition systems.

A new, highly sensitive chemical sensor uses protein nanowires. Scientists report that they have developed bioelectronic ammonia gas sensors that are among the most sensitive ever made. It uses electric-charge-conducting protein nanowires derived from the bacterium Geobacter to provide biomaterials for electrical devices. They grow hair-like protein filaments that work as nanoscale ''wires'' to transfer charges for their nourishment and to communicate with other bacteria.


Motion Sensors - Computer Sensors that Measure Body Language


Motion Detector is a device that detects moving objects, particularly people. Such a device is often integrated as a component of a system that automatically performs a task or alerts a user of motion in an area. They form a vital component of security, automated lighting control, home control, energy efficiency, and other useful systems. TSA Observations.

Algorithm for Analysis of Emotion Using Body Language.

Algorithm to Interpret Body Language in Real Time. Researchers in the United States are working on teaching computers to identify and interpret human physical action sequences as they take place. The algorithm they are using could also be applied to the medical sector to point up correct or incorrect movements by patients.

Hand Body Language Gesture Recognition Based on Signals From Specialized Glove and Machine Learning Algorithms

The Panoptic Studio is a Massively Multiview System, a Social Motion Capture Technology for Recording Body Language and Movements.

Signal Processing is an enabling technology that encompasses the fundamental theory, applications, algorithms, and implementations of processing or transferring information contained in many different physical, symbolic, or abstract formats broadly designated as signals. It uses mathematical, statistical, computational, heuristic, and linguistic representations, formalisms, and techniques for representation, modelling, analysis, synthesis, discovery, recovery, sensing, acquisition, extraction, learning, security, or forensics.

Sensor Fusion is combining of sensory data or data derived from disparate sources such that the resulting information has less uncertainty than would be possible when these sources were used individually. The term uncertainty reduction in this case can mean more accurate, more complete, or more dependable, or refer to the result of an emerging view, such as stereoscopic vision (calculation of depth information by combining two-dimensional images from two cameras at slightly different viewpoints). The data sources for a fusion process are not specified to originate from identical sensors. One can distinguish direct fusion, indirect fusion and fusion of the outputs of the former two. Direct fusion is the fusion of sensor data from a set of heterogeneous or homogeneous sensors, soft sensors, and history values of sensor data, while indirect fusion uses information sources like a priori knowledge about the environment and human input. Sensor fusion is also known as (multi-sensor) data fusion and is a subset of information fusion.



Brain Computer Interface


Do you need a Computer Chip implanted in your Brain?
You don't need a Computer Chip implanted in your Body or Brain, you can still use jump drives, cellphones, paper and other devices to carry important information with you that you need to remember. External Memory Devices are amazing tools. Jumpdrives can be made to look like jewelry or credit cards. You can even get a mini tattoo of your SS# in the form of a QR Code that could link to your information. There are only a few very unique circumstances that would require the need for a person to have a human-implantable microchip, like people who have disabilities. But the body does produce the necessary voltage to run an extra memory device, so I guess it's just a matter of Sensory Substitution. Blending human brains with computers has already been happening for over 20 years, so this is Your Brain on Computers. But you don't need to stick a computer in your head, you can just carry one around in your pocket. You don't have to wait for AI to have enhanced cognitive abilities, you can have enhanced cognitive abilities right now, it's called learning. Human Operating System - Interfaces.

Brain Computer Interface helps Paralyzed Man feel again through Mind-Controlled Robotic Arm.
Primates Regain Control of Paralyzed Limb.
Brain-Computer Interface Laboratory ETSU.
Brain-to-Brain Interface Demonstration (youtube).
Connecting Brains: The BrainNet - (VPRO documentary - 2014) (youtube).
Direct Brain-to-Brain Interface in Humans.
Neuralink is developing implantable brain–computer interfaces (BCIs).
Researchers Revolutionize Brain-Computer Interfaces Using Silicon Electronics.
Mind-Controlled Device helps stroke patients retrain brains to move Paralyzed Hands.
Stroke patient improvement with a brain-computer interface.
Artificial Synapse designed for “Brain-on-a-Chip” Hardware.
Prosthetic memory system successful in humans uses a person’s own memory patterns to facilitate the brain’s ability to encode and recall memory.
Brain-Computer Interface enables people with Paralysis to Control Tablet Devices - Brain Gate.
First-ever successful mind-controlled robotic arm without brain implants using a noninvasive brain-computer interface.
Ultra-small nanoprobes could be a leap forward in high-resolution human-machine interfaces.
Scientists can now manipulate brain cells using smartphone. A team of scientists have invented a device that can control neural circuits using a tiny brain implant controlled by a smartphone.
Controlling electronic devices with brain waves. But researchers have developed a new type of electroencephalogram (EEG) electrode that can do just that, without the sticky gel required for conventional electrodes. Even better, the devices work through a full head of hair. The researchers report the flexible electrodes, which could someday be used in brain-computer interfaces to drive cars or move artificial limbs.
Artificial synapse that works with living cells created. Researchers have created a device that can integrate and interact with neuron-like cells. This could be an early step toward an artificial synapse for use in brain-computer interfaces.

Big Improvements to Brain-Computer Interface. Newly developed “glassy carbon” electrodes transmit more robust signals to restore motion in people with damaged spinal cords.

New treatment allows some people with spinal cord injury to regain hand and arm function. Almost 18,000 Americans experience traumatic spinal cord injuries every year. Many of these people are unable to use their hands and arms and can't do everyday tasks such as eating, grooming or drinking water without help. Using physical therapy combined with a noninvasive method of stimulating nerve cells in the spinal cord, University of Washington researchers helped six Seattle area participants regain some hand and arm mobility. That increased mobility lasted at least three to six months after treatment had ended.

Tapping the Brain to boost Stroke Rehabilitation. Clinical trial suggests brain-machine interface coupled with robot offers increased benefits for stroke survivors. Stroke survivors who had ceased to benefit from conventional rehabilitation gained clinically significant arm movement and control by using an external robotic device or brain-machine interface powered by the patients' own brains.

Human Enhancements will require people to be more educated to avoid the corrupted influences of power. Augmenting.

Body Hacking - Technological Convergence

Brain Computer Interfaces is a direct communication pathway between an enhanced or wired brain and an external device. BCIs are often directed at researching, mapping, assisting, augmenting, or repairing human cognitive or sensory-motor functions.

Brain Computer Interface - Li-Fi - Electrical Brain Stimulation

Stabilizing Brain-Computer interfaces. New machine learning algorithm reduces need for brain-computer interfaces to undergo recalibration.

Neural Engineering System Design program seeks to develop high-resolution neurotechnology capable of mitigating the effects of injury and disease on the visual and auditory systems of military personnel. In addition to creating novel hardware and algorithms, the program conducts research to understand how various forms of neural sensing and actuation might improve restorative therapeutic outcomes. Translator for electrochemical language used by neurons.

Wearable Technology are clothing and accessories incorporating computer and advanced electronic technologies. The designs often incorporate practical functions and features.

Wearable Technology can take basic measurements and monitor and track body functions to give the user a better understand of their body and increase their awareness. They can then match the sensations they feel to the recorded data. Eventually they will be able to teach themselves how to notice body sensations and changes and have a better idea what may be happening in their body. A prosthesis for feeling. And the person will not have to wear the device all the time because they will be more aware of their body sensations and what they may mean. Ai teaching us to be more intelligent, now that's Ai.

Wearable and Bendable Electronics Conference - Medical Sensors

How To Hack A Human Brain | VICE on HBO (youtube) - Advancements in neurotechnology are blurring the line between biology and technology. There is an emerging push to implant electronic devices inside the human body, hardwire them to our brains, and allow us to not only overcome disadvantages or injury but open up entirely new avenues of human experience.

Self-Healing Material are artificial or synthetically-created substances that have the built-in ability to automatically repair damage to themselves without any external diagnosis of the problem or human intervention. Generally, materials will degrade over time due to fatigue, environmental conditions, or damage incurred during operation. Cracks and other types of damage on a microscopic level have been shown to change thermal, electrical, and acoustical properties of materials, and the propagation of cracks can lead to eventual failure of the material.

Bionics is the application of biological methods and systems found in nature to the study and design of engineering systems and modern technology. The Six Million Dollar Man

Artificial Body Parts

Novel nanoprobes show promise for optical monitoring of neural activity. Researchers have developed ultrasensitive nanoscale optical probes to monitor the bioelectric activity of neurons and other excitable cells. This novel readout technology could enable scientists to study how neural circuits function at an unprecedented scale by monitoring large numbers of individual neurons simultaneously. It could also lead to high-bandwidth brain-machine interfaces with dramatically enhanced precision and functionality.

Cyborg is a being with both organic and biomechatronic body parts. Not the same thing as bionic, biorobot or android; it applies to an organism that has restored function or enhanced abilities due to the integration of some artificial component or technology that relies on some sort of feedback. While cyborgs are commonly thought of as mammals, including humans, they might also conceivably be any kind of organism. Augmented.

Cyborg Olympics

Human–Animal Hybrid incorporates elements from both humans and non-human animals.

Cybernetics is a transdisciplinary approach for exploring regulatory systems—their structures, constraints, and possibilities, the scientific study of control and communication in the animal and the machine.

Android is a humanoid robot or synthetic organism designed to look and act like a human, especially one with a body having a flesh-like resemblance.

Humanoid Robot is a robot with its body shape built to resemble the human body. The design may be for functional purposes, such as interacting with human tools and environments, for experimental purposes, such as the study of bipedal locomotion, or for other purposes. In general, humanoid robots have a torso, a head, two arms, and two legs, though some forms of humanoid robots may model only part of the body, for example, from the waist up. Some humanoid robots also have heads designed to replicate human facial features such as eyes and mouths. Androids are humanoid robots built to aesthetically resemble humans.

Robotics - Human Operating System

Brain Implant or Neural Implants, are technological devices that connect directly to a biological subject's brain – usually placed on the surface of the brain, or attached to the brain's cortex. A common purpose of modern brain implants and the focus of much current research is establishing a biomedical prosthesis circumventing areas in the brain that have become dysfunctional after a stroke or other head injuries. This includes sensory substitution, e.g., in vision. Other brain implants are used in animal experiments simply to record brain activity for scientific reasons. Some brain implants involve creating interfaces between neural systems and computer chips. This work is part of a wider research field called brain-computer interfaces. (Brain-computer interface research also includes technology such as EEG arrays that allow interface between mind and machine but do not require direct implantation of a device.) Neural implants such as deep brain stimulation and Vagus nerve stimulation are increasingly becoming routine for patients with Parkinson's disease and clinical depression, respectively.

Man Who Is Paralyzed Communicates By Imagining Handwriting. An experimental device that turns thoughts into text has allowed a man who was left paralyzed by an accident to swiftly construct sentences on a computer screen, able to type with 95% accuracy just by imagining that he was handwriting letters on a sheet of paper. He can type at about 90 characters per minute.

Neuro-Prosthetics is a discipline related to neuroscience and biomedical engineering concerned with developing neural prostheses. They are sometimes contrasted with a brain–computer interface, which connects the brain to a computer rather than a device meant to replace missing biological functionality.

Restoring a rudimentary form of vision in the blind. Recent discoveries show that newly developed high-resolution implants in the visual cortex make it possible to recognize artificially induced shapes and percepts. When electrical stimulation is delivered to the brain via an implanted electrode, it generates the percept of a dot of light at a particular location in visual space, known as a 'phosphene.' The team developed high-resolution implants consisting of 1024 electrodes and implanted them in the visual cortex of two sighted monkeys. Their goal was to create interpretable images by delivering electrical stimulation simultaneously via multiple electrodes, to generate a percept that was composed of multiple phosphenes. "The number of electrodes that we have implanted in the visual cortex, and the number of artificial pixels that we can generate to produce high-resolution artificial images.

Powered Exoskeleton is a wearable mobile machine that is powered by a system of electric motors, pneumatics, levers, hydraulics, or a combination of technologies that allow for limb movement with increased strength and endurance.

Exoskeleton Technology Machines - Wheel Chairs

Hybrid Assistive Limb is a powered exoskeleton suit designed to support and expand the physical capabilities of its users, particularly people with physical disabilities. There are two primary versions of the system: HAL 3, which only provides leg function, and HAL 5, which is a full-body exoskeleton for the arms, legs, and torso.

Thought identification refers to the empirically verified use of technology to, in some sense, read people's minds. Advances in research have made this possible by using human neuroimaging to decode a person's conscious experience based on non-invasive measurements of an individual's brain activity. Repurpose Brain Signals.

RFID or Radio-frequency identification, uses electromagnetic fields to automatically identify and track tags attached to objects. The tags contain electronically stored information. Passive tags collect energy from a nearby RFID reader's interrogating radio waves. Active tags have a local power source such as a battery and may operate at hundreds of meters from the RFID reader. Unlike a barcode, the tag need not be within the line of sight of the reader, so it may be embedded in the tracked object. RFID is one method for Automatic Identification and Data Capture (AIDC)

VeriChip is a human-implantable Microchip, which is an identifying integrated circuit device or RFID transponder encased in silicate glass and implanted in the body of a human being. A subdermal implant typically contains a unique ID number that can be linked to information contained in an external database, such as personal identification, medical history, medications, allergies, and contact information.

Subdermal Implant refers to a body modification that is placed underneath the skin, therefore allowing the body to heal over the implant and creating a raised design. Such implants fall under the broad category of body modification.

Meditation for mind-control. Scientists have discovered that mindful meditation can help subjects learn and improve the ability to mind-control brain computer interfaces (BCIs). A BCI is an apparatus that allows an individual to control a machine or computer directly from their brain. Non-invasive means of control like electroencephalogram (EEG) readings taken through the skull are safe and convenient compared to more risky invasive methods using a brain implant, but they take longer to learn and users ultimately vary in proficiency.

I laugh when I here people say that soon we will be able to upload information directly into our brains, that is so stupid. Why would you need to do that if you have a smart phone or other information storage devices that can carry all your important data with you? And the information you can't carry, you can access it using the internet. Besides you just can't upload information into the brain because information has to be processed very carefully so that the information is correctly understood. So you have to manually and slowly input information into a human brain so that it has time to decipher the information and learn how the information should be used. The key word here is 'Learn', the one word we take for granted. So detailed instructions on how to use this information is a must. Like when you install software into a computer. The software comes with instructions that tells the computer how the information can be used. Then of course the computer needs an operating system in order to use that information correctly. Computers will allow humans to learn faster, but only if the instructions are detailed and accurate. So there are two sides of learning. Choosing the best information to learn and then creating the instructions on how to use the information correctly. Again, the computer and the human brain show their similarities. Reading.

My Favorite Martian was a 1960's sitcom about a Martian stranded on Earth, who claims to be more intelligent than humans, but for some reason, still does stupid human things that contradicts his claim of being more intelligent. Sadly, this show is a reminder of how ignorant society still is, and its been over 50 years. Tracking. My Favorite Martian S3 E06 Tim the Mastermind (youtube) - Taking the pills makes Tim the smartest man on the face of the Earth. (the pills are similar to a brain chip implant). My Favorite Martian S1 E37 " Uncle Martins Wisdom Tooth" (youtube).

Adding Information to our DNA.

We may be limited by our biology, but we are not limited by our intellect. And our intellect has given rise to technology, which is an extension of our intellect. So we have no boundaries. We have taken the first step of controlling matter, which means we have broken through another level of reality, so now what? What is the next step? And why should we take it? It's not the same out there as it is in here, we basically have to start all over again. This is a whole other dimension. Can we live in both dimensions at the same time? And there's word 'Time' again. Time and Space does not have a unified definition, so time and space are like fictitious characters in a play, they're just actors, we know nothing about their personal lives. Even with the mind fully open I still cannot see. So there's the problem, it's not sight, sound, smell, taste or feel. We are so close and yet still so far away. But still an awesome feeling, which way do we go? Lets see how far we can go in both directions without losing contact with each other. And what if we eventually run right back into each other? Then what? 

Can Artificial Intelligence help us find Human Intelligence?

"When people start learning how artificial intelligence can learn on its own, we will actually be teaching ourselves, and at the same time, learn how to learn more effectively."



Previous Subject Up Top Page Next Subject



The Thinker Man