Is the time crystal really an otherworldly revolution, leveraging quantum computing that will change physics forever?
KEY TAKEAWAYS
Google’s quantum computing team, in a first, has developed and demonstrated a discrete time crystal on a quantum computer.
By driving the system with a microwave pulse, they can get it to return to its original quantum state periodically, with no thermal noise or decoherence effects.
However, claims that it violates thermodynamics, is otherworldly, or changes physics forever are all demonstrably untrue and misrepresentative of the actual quality research.
It is tempting, whenever a new discovery comes along, to imagine a whole slew of revolutions that might soon ensue. After all, anytime you can suddenly do or accomplish any task that was previously impractical or even (thought to be) impossible, that is one less obstacle standing in the way of even your loftiest, pie-in-the-sky dreams. However, no matter what discoveries ensue, the fundamental laws of physics that underlie reality must always be obeyed; you might be able to cajole nature into doing a lot of clever things, but you cannot very well break the rules that govern it. If you could, we would have to write down new laws, because the old ones would no longer be valid. Despite all the tricks we have developed, we cannot create fundamental magnetic monopoles, violate the conservation of momentum or energy, or work our way around the second law of thermodynamics.
Yet a recent article, touting a brand new discovery involving time crystals and Google’s quantum computer, claims to do just that: evade the second law of thermodynamics. Is that even right? Patreon supporter Chad Marler wants to know, asking:
Hi Ethan… I was reading some headlines and came across this article. While I suspect the innovation was real, some of the wording in the article made my [nonsense] detector go off… it sounds like something you would hear on an Art Bell show.
I will tell you, up front, that the scientific paper is legit, but the recent article is full of misconceptions and misinterpretations. Let’s dive in and find out what it is all about.
What is a time crystal?
Unlike most things in physics, where theorists imagine a possibility far out of reach of current or near-future technology, time crystals have only been around for a very short time, even in our minds. In 2012, Nobel Laureate Frank Wilczek proposed that a new state of matter might exist: a time crystal. Most of us know what a normal crystal is: a repeating, rigid lattice or grid of particles, like atoms or molecules, that compose a solid, ordered structure.
When we talk about time crystals, however, there is no lattice, no grid, and no solid, ordered structure. The important component of a time crystal, at least conceptually, is the “repeating” part. Whereas a conventional crystal has a structure that repeats in space, meaning it exhibits translational symmetry (if you move within the crystal, the structure looks the same everywhere), a time crystal should repeat its configuration, periodically, in time.
Originally, when time crystals were first considered, they were presumed to be impossible for a number of reasons. There were theorems published that proved their impossibility. There were assertions that a system that transitioned from lower-to-higher energy states would not return to its original state again spontaneously, and then go back-and-forth between those two states, because that would indicate some type of perpetual motion, violating the second law of thermodynamics and the conservation of energy.
But not only did theorists find loopholes in those theorems, but more impressively, experimentalists just went right ahead and created them in the lab. In 2016, Norman Yao and his team came up with a scheme to create a time crystal through a very clever plan. Instead of taking a closed, constant system, he proposed leveraging a system with out-of-equilibrium conditions. He would then “drive” that system externally, making it an open (rather than a closed) system and achieving the much sought after “time crystal” state.
It is a little bit complicated, but you can imagine that you have a bunch of atoms that have a spin, and those spins have directions: dipole moments. The way you can “drive” the system is by subjecting the system to spin-echo pulses that contain imperfections, but which occur periodically while allowing interactions to randomly occur in the intermediate times. If you get the combinations of these dipole moments of the spins and the spin-echo pulses to behave in a certain fashion, you could get a time crystal.
The hard part, though, is avoiding what normally happens when you interact with a system: If there is an exchange of energy, that energy gets transferred throughout the system, internally, causing runaway heating due to many-body interactions. Somehow, you have to:
drive the system, externally, with a spin-flip pulse,
so that you get a periodic response,
that is proportional to the time at which you pulse the system,
and at some multiple of the period, you return to your initial state,
while the “time crystal” only oscillates away from and then back into that initial state.
Only if you go back, periodically, to exactly your initial state, with no extra heating and achieve a pure steady-state can you make a time crystal.
They tried to set up a system precisely as Yao had demanded but, because the conditions are so general, wound up taking vastly different approaches.
Monroe’s group took a series of yttrium atoms all lined up, in a one-dimensional line, all coupled together via their electrostatic interactions. When they subjected this atomic line to a series of spin-flip pulses, they found that the system would return to its initial state every two full pulse periods. Meanwhile, Lukin’s group took an actual diamond crystal that contained somewhere on the order of ~1,000,000 spin-impurities within it and pulsed those impurities within the crystal with microwave radiation. That radiation flipped their spins, but time crystal oscillations were only observed every three full pulse periods — whereupon the crystal would return to its initial quantum state.
This occurred for both groups, interestingly enough, even when the driving pulses were imperfect. You could:
alter the magnitude of the pulse, making it stronger or weaker,
vary the frequency of pulsation, making it a little quicker or slower,
turn up or turn down the amount of noise and/or interactions that occurred between the pulses,
or change the conditions of the environment that the system is placed in,
and still recover this time crystal behavior. Surprisingly, for these non-equilibrium systems, there is a lot of wiggle-room as far as what you can do and still observe this time crystal behavior.
But as they were originally envisioned by Wilczek in 2012, an idealized time crystal would occur in a system that was in thermal equilibrium — that was neither absorbing nor emitting energy from or to the surrounding environment. In order to create a time crystal, you needed to have an open system that could exchange energy with its external surroundings, and that system needed to be driven at a periodic frequency. Moreover, the imperfections in the driving could not be too large, or the crystal would “melt” in precisely the fashion we want to avoid: with runaway heating occurring from many-body interactions.
What did the Google team, using a quantum computer, actually do?
Back when these time crystals were first realized in 2016/2017, it was recognized that time crystals could conceivably be applied to quantum computers. Instead of encoding a bit, like the “0” or “1” a standard computer encodes, a quantum computer encodes a qubit, which is a probability-weighted superposition of both “0” and “1” simultaneously. Although you can only measure a “0” or “1” at the end, the fact that you have many qubits allows you to see whether you have preserved the quantum behavior of the system (or not), whether your results are error-free (or not), and what type of final-state distribution you get and whether it matches your theoretical predictions.
The hard “problem” with a quantum computer is the problem of decoherence: Over relatively short timescales, the system interacts with the surrounding particles, and this causes you to lose the quantum behavior you are trying to preserve. For Google’s quantum computer, which is based on superconducting qubits (as opposed to quantum dots or ion traps, for example), you get a coherence timescale of about 50 microseconds. You can only perform perhaps a few dozen computations before decoherence ruins your experiment, and you lose the quantum behavior you sought to maintain and measure. (Or, more precisely, before too many errors, including errors from simple crosstalk between qubits, simply transform your signal into noise.)
Instead of using a dynamical phase like the spins of atoms, though, a quantum computer allows you to use a different property: the order of eigenstates in many-body systems. If you brought your qubits into an equilibrium setting, you would see that there was order in the lowest energy states and unordered states at higher energies. That is why, under normal circumstances, if you allow too much energy to propagate through your system, you just wind up with featureless, unordered systems; it is like the heat or energy just randomized everything.
However, some systems can exhibit what is called MBL: many-body localization, where you get local conservation laws and only a discrete number of ordered states. When you drive the system, which the Google team did with pulsed microwaves that cause the qubits to flip, your qubits have the potential to behave just like the dynamical phases did when we were measuring atomic spins: If the qubits do not absorb heat or impart energy to their surroundings, they can simply flip between different ordered states. With enough pulses, you can conceivably recover your original state.
Sure enough, every two full periods of the microwave pulses resulted in a recovery of the original state: a time crystal. Not bound by these decoherence effects any longer, the researchers could maintain this time crystal state for up to ~100 seconds, a remarkable achievement.
Although the article does a fine job of describing the experiments performed themselves, there is a howler of a statement made early on:
With the ability to forever cycle between two states without ever losing energy, time crystals dodge one of the most important laws of physics — the second law of thermodynamics, which states that the disorder, or entropy, of an isolated system must always increase. These bizarre time crystals remain stable, resisting any dissolution into randomness, despite existing in a constant state of flux.
There is no dodge; the second law of thermodynamics applies to closed systems, not open ones. The disorder of the system, if you include the microwave pulses and the external environment, does in fact go up, just as predicted. The crystals oscillate between allowable states and return to their original ones when driven properly, just as their non-qubit analogues did years prior. In order to do this, the researchers needed to discriminate between external decoherence and internal thermalization, both of which can destroy the quantum state they are seeking to maintain, which itself is an admirable achievement.
Although it may be fun to claim, as the headline of the article did, that this is “otherworldly” and “could change physics forever,” it is more like imagining you have got a skillet with different sized and shaped mollusks in it and a chef who jiggles the pan in a way that makes the shelled creatures flip. Some will flip 180°, others 360°, others 540°, etc. In the quantum world, some of these mollusks can take on in-between values, too. But after a certain number of jiggles, the mollusks all wind up the same way they started, regardless of what that specific initial configuration was. That is all the Google team is doing, but instead of mollusks or spinning atoms, they are using the eigenstates of a quantum computer.
Which, if we are being honest, is still a remarkable achievement! This is a new kind of time crystal, a new way of achieving it, and one with the potential to study non-equilibrium phases of matter on a quantum computer. And although you have to pump energy into the system in pulses, the time crystal can, in fact, return to whatever specific state it began with, even with small imperfections occurring in the “flips,” without destroying, decohering, or losing the nature of the quantum state due to thermal instabilities. No laws are violated and the physics we know is not changed in any way, but this is a phenomenal achievement nonetheless. In a mere nine years, we have gone from theorizing the existence of time crystals to creating them to observing them on a quantum processor. When a new field yields significant advances so quickly, it compels us to pay attention.
Information Security Attributes: or qualities, i.e., Confidentiality, Integrity and Availability (CIA). Information Systems are composed in three main portions, hardware, software and communications with the purpose to help identify and apply information security industry standards, as mechanisms of protection and prevention, at three levels or layers: physical, personal and organizational. Essentially, procedures or policies are implemented to tell administrators, users and operators how to use products to ensure information security within the organizations.[10]
Various definitions of information security are suggested below, summarized from different sources:
“Preservation of confidentiality, integrity and availability of information. Note: In addition, other properties, such as authenticity, accountability, non-repudiation and reliability can also be involved.” (ISO/IEC 27000:2009)[11]
“The protection of information and information systems from unauthorized access, use, disclosure, disruption, modification, or destruction in order to provide confidentiality, integrity, and availability.” (CNSS, 2010)[12]
“Ensures that only authorized users (confidentiality) have access to accurate and complete information (integrity) when required (availability).” (ISACA, 2008)[13]
“Information Security is the process of protecting the intellectual property of an organisation.” (Pipkin, 2000)[14]
“…information security is a risk management discipline, whose job is to manage the cost of information risk to the business.” (McDermott and Geer, 2001)[15]
“A well-informed sense of assurance that information risks and controls are in balance.” (Anderson, J., 2003)[16]
“Information security is the protection of information and minimizes the risk of exposing information to unauthorized parties.” (Venter and Eloff, 2003)[17]
“Information Security is a multidisciplinary area of study and professional activity which is concerned with the development and implementation of security mechanisms of all available types (technical, organizational, human-oriented and legal) in order to keep information in all its locations (within and outside the organization’s perimeter) and, consequently, information systems, where information is created, processed, stored, transmitted and destroyed, free from threats.[18] Threats to information and information systems may be categorized and a corresponding security goal may be defined for each category of threats.[19] A set of security goals, identified as a result of a threat analysis, should be revised periodically to ensure its adequacy and conformance with the evolving environment.[20] The currently relevant set of security goals may include: confidentiality, integrity, availability, privacy, authenticity & trustworthiness, non-repudiation, accountability and auditability.” (Cherdantseva and Hilton, 2013)[10]
Information and information resource security using telecommunication system or devices means protecting information, information systems or books from unauthorized access, damage, theft, or destruction (Kurose and Ross, 2010).[21]
Overview
At the core of information security is information assurance, the act of maintaining the confidentiality, integrity, and availability (CIA) of information, ensuring that information is not compromised in any way when critical issues arise.[22] These issues include but are not limited to natural disasters, computer/server malfunction, and physical theft. While paper-based business operations are still prevalent, requiring their own set of information security practices, enterprise digital initiatives are increasingly being emphasized,[23][24] with information assurance now typically being dealt with by information technology (IT) security specialists. These specialists apply information security to technology (most often some form of computer system). It is worthwhile to note that a computer does not necessarily mean a home desktop.[25] A computer is any device with a processor and some memory. Such devices can range from non-networked standalone devices as simple as calculators, to networked mobile computing devices such as smartphones and tablet computers.[26] IT security specialists are almost always found in any major enterprise/establishment due to the nature and value of the data within larger businesses.[27] They are responsible for keeping all of the technology within the company secure from malicious cyber attacks that often attempt to acquire critical private information or gain control of the internal systems.[28][29]
The field of information security has grown and evolved significantly in recent years.[30] It offers many areas for specialization, including securing networks and allied infrastructure, securing applications and databases, security testing, information systems auditing, business continuity planning, electronic record discovery, and digital forensics.[citation needed] Information security professionals are very stable in their employment.[31] As of 2013 more than 80 percent of professionals had no change in employer or employment over a period of a year, and the number of professionals is projected to continuously grow more than 11 percent annually from 2014 to 2019.[32]
Threats
Information security threats come in many different forms.[33][34] Some of the most common threats today are software attacks, theft of intellectual property, theft of identity, theft of equipment or information, sabotage, and information extortion.[35][36] Most people have experienced software attacks of some sort. Viruses,[37]worms, phishing attacks, and Trojan horses are a few common examples of software attacks. The theft of intellectual property has also been an extensive issue for many businesses in the information technology (IT) field.[38]Identity theft is the attempt to act as someone else usually to obtain that person’s personal information or to take advantage of their access to vital information through social engineering.[39][40] Theft of equipment or information is becoming more prevalent today due to the fact that most devices today are mobile,[41] are prone to theft and have also become far more desirable as the amount of data capacity increases. Sabotage usually consists of the destruction of an organization’s website in an attempt to cause loss of confidence on the part of its customers.[42] Information extortion consists of theft of a company’s property or information as an attempt to receive a payment in exchange for returning the information or property back to its owner, as with ransomware.[43] There are many ways to help protect yourself from some of these attacks but one of the most functional precautions is conduct periodical user awareness.[44] The number one threat to any organisation are users or internal employees, they are also called insider threats.[45]
Governments, military, corporations, financial institutions, hospitals, non-profit organisations, and private businesses amass a great deal of confidential information about their employees, customers, products, research, and financial status.[46] Should confidential information about a business’ customers or finances or new product line fall into the hands of a competitor or a black hat hacker, a business and its customers could suffer widespread, irreparable financial loss, as well as damage to the company’s reputation.[47] From a business perspective, information security must be balanced against cost; the Gordon-Loeb Model provides a mathematical economic approach for addressing this concern.[48]
For the individual, information security has a significant effect on privacy, which is viewed very differently in various cultures.[49]
Responses to threats
Possible responses to a security threat or risk are:[50]
reduce/mitigate – implement safeguards and countermeasures to eliminate vulnerabilities or block threats
assign/transfer – place the cost of the threat onto another entity or organization such as purchasing insurance or outsourcing
accept – evaluate if the cost of the countermeasure outweighs the possible cost of loss due to the threat[51]
History
Since the early days of communication, diplomats and military commanders understood that it was necessary to provide some mechanism to protect the confidentiality of correspondence and to have some means of detecting tampering.[52]Julius Caesar is credited with the invention of the Caesar cipher c. 50 B.C., which was created in order to prevent his secret messages from being read should a message fall into the wrong hands.[53] However, for the most part protection was achieved through the application of procedural handling controls.[54][55] Sensitive information was marked up to indicate that it should be protected and transported by trusted persons, guarded and stored in a secure environment or strong box.[56] As postal services expanded, governments created official organizations to intercept, decipher, read, and reseal letters (e.g., the U.K.’s Secret Office, founded in 1653[57]).
In the mid-nineteenth century more complex classification systems were developed to allow governments to manage their information according to the degree of sensitivity.[58] For example, the British Government codified this, to some extent, with the publication of the Official Secrets Act in 1889.[59] Section 1 of the law concerned espionage and unlawful disclosures of information, while Section 2 dealt with breaches of official trust.[60] A public interest defense was soon added to defend disclosures in the interest of the state.[61] A similar law was passed in India in 1889, The Indian Official Secrets Act, which was associated with the British colonial era and used to crack down on newspapers that opposed the Raj’s policies.[62] A newer version was passed in 1923 that extended to all matters of confidential or secret information for governance.[63] By the time of the First World War, multi-tier classification systems were used to communicate information to and from various fronts, which encouraged greater use of code making and breaking sections in diplomatic and military headquarters.[64] Encoding became more sophisticated between the wars as machines were employed to scramble and unscramble information.[65]
The establishment of computer security inaugurated the history of information security. The need for such appeared during World War II.[66] The volume of information shared by the Allied countries during the Second World War necessitated formal alignment of classification systems and procedural controls.[67] An arcane range of markings evolved to indicate who could handle documents (usually officers rather than enlisted troops) and where they should be stored as increasingly complex safes and storage facilities were developed.[68] The Enigma Machine, which was employed by the Germans to encrypt the data of warfare and was successfully decrypted by Alan Turing, can be regarded as a striking example of creating and using secured information.[69] Procedures evolved to ensure documents were destroyed properly, and it was the failure to follow these procedures which led to some of the greatest intelligence coups of the war (e.g., the capture of U-570[69]).
In 1973, important elements of ARPANET security were found by internet pioneer Robert Metcalfe to have many flaws such as the: “vulnerability of password structure and formats; lack of safety procedures for dial-up connections; and nonexistent user identification and authorizations”, aside from the lack of controls and safeguards to keep data safe from unauthorized access. Hackers had effortless access to ARPANET, as phone numbers were known by the public.[71] Due to these problems, coupled with the constant violation of computer security, as well as the exponential increase in the number of hosts and users of the system, “network security” was often alluded to as “network insecurity”.[71]
The end of the twentieth century and the early years of the twenty-first century saw rapid advancements in telecommunications, computing hardware and software, and data encryption.[72] The availability of smaller, more powerful, and less expensive computing equipment made electronic data processing within the reach of small business and home users.[73] The establishment of Transfer Control Protocol/Internetwork Protocol (TCP/IP) in the early 1980s enabled different types of computers to communicate.[74] These computers quickly became interconnected through the internet.[75]
The rapid growth and widespread use of electronic data processing and electronic business conducted through the internet, along with numerous occurrences of international terrorism, fueled the need for better methods of protecting the computers and the information they store, process, and transmit.[76] The academic disciplines of computer security and information assurance emerged along with numerous professional organizations, all sharing the common goals of ensuring the security and reliability of information systems.[citation needed]
The CIA triad of confidentiality, integrity, and availability is at the heart of information security.[77] (The members of the classic InfoSec triad—confidentiality, integrity, and availability—are interchangeably referred to in the literature as security attributes, properties, security goals, fundamental aspects, information criteria, critical information characteristics and basic building blocks.)[78] However, debate continues about whether or not this CIA triad is sufficient to address rapidly changing technology and business requirements, with recommendations to consider expanding on the intersections between availability and confidentiality, as well as the relationship between security and privacy.[22] Other principles such as “accountability” have sometimes been proposed; it has been pointed out that issues such as non-repudiation do not fit well within the three core concepts.[79]
The triad seems to have first been mentioned in a NIST publication in 1977.[80]
In 1992 and revised in 2002, the OECD‘s Guidelines for the Security of Information Systems and Networks[81] proposed the nine generally accepted principles: awareness, responsibility, response, ethics, democracy, risk assessment, security design and implementation, security management, and reassessment.[82] Building upon those, in 2004 the NIST‘s Engineering Principles for Information Technology Security[79] proposed 33 principles. From each of these derived guidelines and practices.
In information security, confidentiality “is the property, that information is not made available or disclosed to unauthorized individuals, entities, or processes.”[88] While similar to “privacy,” the two words aren’t interchangeable. Rather, confidentiality is a component of privacy that implements to protect our data from unauthorized viewers.[89] Examples of confidentiality of electronic data being compromised include laptop theft, password theft, or sensitive emails being sent to the incorrect individuals.[90]
Integrity
In IT security, data integrity means maintaining and assuring the accuracy and completeness of data over its entire lifecycle.[91] This means that data cannot be modified in an unauthorized or undetected manner.[92] This is not the same thing as referential integrity in databases, although it can be viewed as a special case of consistency as understood in the classic ACID model of transaction processing.[93] Information security systems typically incorporate controls to ensure their own integrity, in particular protecting the kernel or core functions against both deliberate and accidental threats.[94] Multi-purpose and multi-user computer systems aim to compartmentalize the data and processing such that no user or process can adversely impact another: the controls may not succeed however, as we see in incidents such as malware infections, hacks, data theft, fraud, and privacy breaches.[95]
More broadly, integrity is an information security principle that involves human/social, process, and commercial integrity, as well as data integrity. As such it touches on aspects such as credibility, consistency, truthfulness, completeness, accuracy, timeliness, and assurance.[96]
Availability
For any information system to serve its purpose, the information must be available when it is needed.[97] This means the computing systems used to store and process the information, the security controls used to protect it, and the communication channels used to access it must be functioning correctly.[98]High availability systems aim to remain available at all times, preventing service disruptions due to power outages, hardware failures, and system upgrades.[99] Ensuring availability also involves preventing denial-of-service attacks, such as a flood of incoming messages to the target system, essentially forcing it to shut down.[100]
In the realm of information security, availability can often be viewed as one of the most important parts of a successful information security program.[citation needed] Ultimately end-users need to be able to perform job functions; by ensuring availability an organization is able to perform to the standards that an organization’s stakeholders expect.[101] This can involve topics such as proxy configurations, outside web access, the ability to access shared drives and the ability to send emails.[102] Executives oftentimes do not understand the technical side of information security and look at availability as an easy fix, but this often requires collaboration from many different organizational teams, such as network operations, development operations, incident response, and policy/change management.[103] A successful information security team involves many different key roles to mesh and align for the CIA triad to be provided effectively.[104]
Non-repudiation
In law, non-repudiation implies one’s intention to fulfill their obligations to a contract. It also implies that one party of a transaction cannot deny having received a transaction, nor can the other party deny having sent a transaction.[105]
It is important to note that while technology such as cryptographic systems can assist in non-repudiation efforts, the concept is at its core a legal concept transcending the realm of technology.[106] It is not, for instance, sufficient to show that the message matches a digital signature signed with the sender’s private key, and thus only the sender could have sent the message, and nobody else could have altered it in transit (data integrity).[107] The alleged sender could in return demonstrate that the digital signature algorithm is vulnerable or flawed, or allege or prove that his signing key has been compromised.[108] The fault for these violations may or may not lie with the sender, and such assertions may or may not relieve the sender of liability, but the assertion would invalidate the claim that the signature necessarily proves authenticity and integrity. As such, the sender may repudiate the message (because authenticity and integrity are pre-requisites for non-repudiation).[109]
Broadly speaking, risk is the likelihood that something bad will happen that causes harm to an informational asset (or the loss of the asset).[110] A vulnerability is a weakness that could be used to endanger or cause harm to an informational asset. A threat is anything (man-made or act of nature) that has the potential to cause harm.[111] The likelihood that a threat will use a vulnerability to cause harm creates a risk. When a threat does use a vulnerability to inflict harm, it has an impact.[112] In the context of information security, the impact is a loss of availability, integrity, and confidentiality, and possibly other losses (lost income, loss of life, loss of real property).[113]
The Certified Information Systems Auditor (CISA) Review Manual 2006 defines risk management as “the process of identifying vulnerabilities and threats to the information resources used by an organization in achieving business objectives, and deciding what countermeasures,[114] if any, to take in reducing risk to an acceptable level, based on the value of the information resource to the organization.”[115]
There are two things in this definition that may need some clarification. First, the process of risk management is an ongoing, iterative process. It must be repeated indefinitely. The business environment is constantly changing and new threats and vulnerabilities emerge every day.[116] Second, the choice of countermeasures (controls) used to manage risks must strike a balance between productivity, cost, effectiveness of the countermeasure, and the value of the informational asset being protected.[117] Furthermore, these processes have limitations as security breaches are generally rare and emerge in a specific context which may not be easily duplicated.[118] Thus, any process and countermeasure should itself be evaluated for vulnerabilities.[119] It is not possible to identify all risks, nor is it possible to eliminate all risk. The remaining risk is called “residual risk.[120]“
A risk assessment is carried out by a team of people who have knowledge of specific areas of the business.[121] Membership of the team may vary over time as different parts of the business are assessed.[122] The assessment may use a subjective qualitative analysis based on informed opinion, or where reliable dollar figures and historical information is available, the analysis may use quantitative analysis.
Research has shown that the most vulnerable point in most information systems is the human user, operator, designer, or other human.[123] The ISO/IEC 27002:2005 Code of practice for information security management recommends the following be examined during a risk assessment:
Calculate the impact that each threat would have on each asset. Use qualitative analysis or quantitative analysis.[129]
Identify, select and implement appropriate controls. Provide a proportional response. Consider productivity, cost effectiveness, and value of the asset.[130]
Evaluate the effectiveness of the control measures. Ensure the controls provide the required cost effective protection without discernible loss of productivity.[131]
For any given risk, management can choose to accept the risk based upon the relative low value of the asset, the relative low frequency of occurrence, and the relative low impact on the business.[132] Or, leadership may choose to mitigate the risk by selecting and implementing appropriate control measures to reduce the risk. In some cases, the risk can be transferred to another business by buying insurance or outsourcing to another business.[133] The reality of some risks may be disputed. In such cases leadership may choose to deny the risk.[134]
Selecting and implementing proper security controls will initially help an organization bring down risk to acceptable levels.[135] Control selection should follow and should be based on the risk assessment.[136] Controls can vary in nature, but fundamentally they are ways of protecting the confidentiality, integrity or availability of information. ISO/IEC 27001 has defined controls in different areas.[137] Organizations can implement additional controls according to requirement of the organization.[138]ISO/IEC 27002 offers a guideline for organizational information security standards.[139]
Administrative
Administrative controls (also called procedural controls) consist of approved written policies, procedures, standards, and guidelines. Administrative controls form the framework for running the business and managing people.[140] They inform people on how the business is to be run and how day-to-day operations are to be conducted. Laws and regulations created by government bodies are also a type of administrative control because they inform the business.[141] Some industry sectors have policies, procedures, standards, and guidelines that must be followed – the Payment Card Industry Data Security Standard[142] (PCI DSS) required by Visa and MasterCard is such an example. Other examples of administrative controls include the corporate security policy, password policy, hiring policies, and disciplinary policies.[143]
Administrative controls form the basis for the selection and implementation of logical and physical controls. Logical and physical controls are manifestations of administrative controls, which are of paramount importance.[140]
Logical
Logical controls (also called technical controls) use software and data to monitor and control access to information and computing systems.[citation needed] Passwords, network and host-based firewalls, network intrusion detection systems, access control lists, and data encryption are examples of logical controls.[144]
An important logical control that is frequently overlooked is the principle of least privilege, which requires that an individual, program or system process not be granted any more access privileges than are necessary to perform the task.[145] A blatant example of the failure to adhere to the principle of least privilege is logging into Windows as user Administrator to read email and surf the web. Violations of this principle can also occur when an individual collects additional access privileges over time.[146] This happens when employees’ job duties change, employees are promoted to a new position, or employees are transferred to another department.[147] The access privileges required by their new duties are frequently added onto their already existing access privileges, which may no longer be necessary or appropriate.[148]
Physical
Physical controls monitor and control the environment of the work place and computing facilities.[149] They also monitor and control access to and from such facilities and include doors, locks, heating and air conditioning, smoke and fire alarms, fire suppression systems, cameras, barricades, fencing, security guards, cable locks, etc. Separating the network and workplace into functional areas are also physical controls.[150]
An important physical control that is frequently overlooked is separation of duties, which ensures that an individual can not complete a critical task by himself.[151] For example, an employee who submits a request for reimbursement should not also be able to authorize payment or print the check.[152] An applications programmer should not also be the server administrator or the database administrator; these roles and responsibilities must be separated from one another.[153]
Information security must protect information throughout its lifespan, from the initial creation of the information on through to the final disposal of the information.[154] The information must be protected while in motion and while at rest. During its lifetime, information may pass through many different information processing systems and through many different parts of information processing systems.[155] There are many different ways the information and information systems can be threatened. To fully protect the information during its lifetime, each component of the information processing system must have its own protection mechanisms.[156] The building up, layering on, and overlapping of security measures is called “defense in depth.”[157] In contrast to a metal chain, which is famously only as strong as its weakest link, the defense in depth strategy aims at a structure where, should one defensive measure fail, other measures will continue to provide protection.[158]
Recall the earlier discussion about administrative controls, logical controls, and physical controls. The three types of controls can be used to form the basis upon which to build a defense in depth strategy.[140] With this approach, defense in depth can be conceptualized as three distinct layers or planes laid one on top of the other.[159] Additional insight into defense in depth can be gained by thinking of it as forming the layers of an onion, with data at the core of the onion, people the next outer layer of the onion, and network security, host-based security, and application security forming the outermost layers of the onion.[160] Both perspectives are equally valid, and each provides valuable insight into the implementation of a good defense in depth strategy.[161]
An important aspect of information security and risk management is recognizing the value of information and defining appropriate procedures and protection requirements for the information.[162] Not all information is equal and so not all information requires the same degree of protection.[163] This requires information to be assigned a security classification.[164] The first step in information classification is to identify a member of senior management as the owner of the particular information to be classified. Next, develop a classification policy.[165] The policy should describe the different classification labels, define the criteria for information to be assigned a particular label, and list the required security controls for each classification.[166]
Some factors that influence which classification information should be assigned include how much value that information has to the organization, how old the information is and whether or not the information has become obsolete.[167] Laws and other regulatory requirements are also important considerations when classifying information.[168] The Information Systems Audit and Control Association (ISACA) and its Business Model for Information Security also serves as a tool for security professionals to examine security from a systems perspective, creating an environment where security can be managed holistically, allowing actual risks to be addressed.[169]
The type of information security classification labels selected and used will depend on the nature of the organization, with examples being:[166]
In the business sector, labels such as: Public, Sensitive, Private, Confidential.
In the government sector, labels such as: Unclassified, Unofficial, Protected, Confidential, Secret, Top Secret, and their non-English equivalents.[170]
In cross-sectoral formations, the Traffic Light Protocol, which consists of: White, Green, Amber, and Red.
All employees in the organization, as well as business partners, must be trained on the classification schema and understand the required security controls and handling procedures for each classification.[171] The classification of a particular information asset that has been assigned should be reviewed periodically to ensure the classification is still appropriate for the information and to ensure the security controls required by the classification are in place and are followed in their right procedures.[172]
Access control
Access to protected information must be restricted to people who are authorized to access the information.[173] The computer programs, and in many cases the computers that process the information, must also be authorized.[174] This requires that mechanisms be in place to control the access to protected information.[174] The sophistication of the access control mechanisms should be in parity with the value of the information being protected; the more sensitive or valuable the information the stronger the control mechanisms need to be.[175] The foundation on which access control mechanisms are built start with identification and authentication.[176]
Identification is an assertion of who someone is or what something is. If a person makes the statement “Hello, my name is John Doe” they are making a claim of who they are.[178] However, their claim may or may not be true. Before John Doe can be granted access to protected information it will be necessary to verify that the person claiming to be John Doe really is John Doe.[179] Typically the claim is in the form of a username. By entering that username you are claiming “I am the person the username belongs to”.[180]
Authentication
Authentication is the act of verifying a claim of identity. When John Doe goes into a bank to make a withdrawal, he tells the bank teller he is John Doe, a claim of identity.[181] The bank teller asks to see a photo ID, so he hands the teller his driver’s license.[182] The bank teller checks the license to make sure it has John Doe printed on it and compares the photograph on the license against the person claiming to be John Doe.[183] If the photo and name match the person, then the teller has authenticated that John Doe is who he claimed to be. Similarly, by entering the correct password, the user is providing evidence that he/she is the person the username belongs to.[184]
There are three different types of information that can be used for authentication:[185][186]
Strong authentication requires providing more than one type of authentication information (two-factor authentication).[192] The username is the most common form of identification on computer systems today and the password is the most common form of authentication.[193] Usernames and passwords have served their purpose, but they are increasingly inadequate.[194] Usernames and passwords are slowly being replaced or supplemented with more sophisticated authentication mechanisms such as Time-based One-time Password algorithms.[195]
Authorization
After a person, program or computer has successfully been identified and authenticated then it must be determined what informational resources they are permitted to access and what actions they will be allowed to perform (run, view, create, delete, or change).[196] This is called authorization. Authorization to access information and other computing services begins with administrative policies and procedures.[197] The policies prescribe what information and computing services can be accessed, by whom, and under what conditions. The access control mechanisms are then configured to enforce these policies.[198] Different computing systems are equipped with different kinds of access control mechanisms. Some may even offer a choice of different access control mechanisms.[199] The access control mechanism a system offers will be based upon one of three approaches to access control, or it may be derived from a combination of the three approaches.[90]
The non-discretionary approach consolidates all access control under a centralized administration.[200] The access to information and other resources is usually based on the individuals function (role) in the organization or the tasks the individual must perform.[201][202] The discretionary approach gives the creator or owner of the information resource the ability to control access to those resources.[200] In the mandatory access control approach, access is granted or denied basing upon the security classification assigned to the information resource.[173]
To be effective, policies and other security controls must be enforceable and upheld. Effective policies ensure that people are held accountable for their actions.[205] The U.S. Treasury‘s guidelines for systems processing sensitive or proprietary information, for example, states that all failed and successful authentication and access attempts must be logged, and all access to information must leave some type of audit trail.[206]
Also, the need-to-know principle needs to be in effect when talking about access control. This principle gives access rights to a person to perform their job functions.[207] This principle is used in the government when dealing with difference clearances.[208] Even though two employees in different departments have a top-secret clearance, they must have a need-to-know in order for information to be exchanged. Within the need-to-know principle, network administrators grant the employee the least amount of privilege to prevent employees from accessing more than what they are supposed to.[209] Need-to-know helps to enforce the confidentiality-integrity-availability triad. Need-to-know directly impacts the confidential area of the triad.[210]
Information security uses cryptography to transform usable information into a form that renders it unusable by anyone other than an authorized user; this process is called encryption.[211] Information that has been encrypted (rendered unusable) can be transformed back into its original usable form by an authorized user who possesses the cryptographic key, through the process of decryption.[212] Cryptography is used in information security to protect information from unauthorized or accidental disclosure while the information is in transit (either electronically or physically) and while information is in storage.[90]
Cryptography provides information security with other useful applications as well, including improved authentication methods, message digests, digital signatures, non-repudiation, and encrypted network communications.[213] Older, less secure applications such as Telnet and File Transfer Protocol (FTP) are slowly being replaced with more secure applications such as Secure Shell (SSH) that use encrypted network communications.[214] Wireless communications can be encrypted using protocols such as WPA/WPA2 or the older (and less secure) WEP. Wired communications (such as ITU‑TG.hn) are secured using AES for encryption and X.1035 for authentication and key exchange.[215] Software applications such as GnuPG or PGP can be used to encrypt data files and email.[216]
Cryptography can introduce security problems when it is not implemented correctly.[217] Cryptographic solutions need to be implemented using industry-accepted solutions that have undergone rigorous peer review by independent experts in cryptography.[218] The length and strength of the encryption key is also an important consideration.[219] A key that is weak or too short will produce weak encryption.[219] The keys used for encryption and decryption must be protected with the same degree of rigor as any other confidential information.[220] They must be protected from unauthorized disclosure and destruction, and they must be available when needed.[221]Public key infrastructure (PKI) solutions address many of the problems that surround key management.[90]
Process
The terms “reasonable and prudent person”, “due care“, and “due diligence” have been used in the fields of finance, securities, and law for many years. In recent years these terms have found their way into the fields of computing and information security.[125] U.S. Federal Sentencing Guidelines now make it possible to hold corporate officers liable for failing to exercise due care and due diligence in the management of their information systems.[222]
In the business world, stockholders, customers, business partners, and governments have the expectation that corporate officers will run the business in accordance with accepted business practices and in compliance with laws and other regulatory requirements. This is often described as the “reasonable and prudent person” rule. A prudent person takes due care to ensure that everything necessary is done to operate the business by sound business principles and in a legal, ethical manner. A prudent person is also diligent (mindful, attentive, ongoing) in their due care of the business.
In the field of information security, Harris[223] offers the following definitions of due care and due diligence:
“Due care are steps that are taken to show that a company has taken responsibility for the activities that take place within the corporation and has taken the necessary steps to help protect the company, its resources, and employees[224].” And, [Due diligence are the] “continual activities that make sure the protection mechanisms are continually maintained and operational.”[225]
Attention should be made to two important points in these definitions.[226][227] First, in due care, steps are taken to show; this means that the steps can be verified, measured, or even produce tangible artifacts.[228][229] Second, in due diligence, there are continual activities; this means that people are actually doing things to monitor and maintain the protection mechanisms, and these activities are ongoing.[230]
Organizations have a responsibility with practicing duty of care when applying information security. The Duty of Care Risk Analysis Standard (DoCRA)[231] provides principles and practices for evaluating risk.[232] It considers all parties that could be affected by those risks.[233] DoCRA helps evaluate safeguards if they are appropriate in protecting others from harm while presenting a reasonable burden.[234] With increased data breach litigation, companies must balance security controls, compliance, and its mission.[235]
Roles, responsibilities, and segregation of duties defined
Addressed and enforced in policy
Adequate resources committed
Staff aware and trained
A development life cycle requirement
Planned, managed, measurable, and measured
Reviewed and audited
Incident response plans
This section needs expansion. You can help by adding to it. (January 2018)
An incident response plan (IRP) is a group of policies that dictate an organizations reaction to a cyber attack. Once an security breach has been identified the plan is initiated.[237] It is important to note that there can be legal implications to a data breach. Knowing local and federal laws is critical.[238] Every plan is unique to the needs of the organization, and it can involve skill sets that are not part of an IT team.[239] For example, a lawyer may be included in the response plan to help navigate legal implications to a data breach.[240]
As mentioned above every plan is unique but most plans will include the following:[241]
Preparation
Good preparation includes the development of an Incident Response Team (IRT).[242] Skills need to be used by this team would be, penetration testing, computer forensics, network security, etc.[243] This team should also keep track of trends in cybersecurity and modern attack strategies.[244] A training program for end users is important as well as most modern attack strategies target users on the network.[241]
Identification
This part of the incident response plan identifies if there was a security event.[245] When an end user reports information or an admin notices irregularities, an investigation is launched. An incident log is a crucial part of this step.[246] All of the members of the team should be updating this log to ensure that information flows as fast as possible.[247] If it has been identified that a security breach has occurred the next step should be activated.[248]
Containment
In this phase, the IRT works to isolate the areas that the breach took place to limit the scope of the security event.[249] During this phase it is important to preserve information forensically so it can be analyzed later in the process.[250] Containment could be as simple as physically containing a server room or as complex as segmenting a network to not allow the spread of a virus.[251]
Eradication
This is where the threat that was identified is removed from the affected systems.[252] This could include using deleting malicious files, terminating compromised accounts, or deleting other components.[253][254] Some events do not require this step, however it is important to fully understand the event before moving to this step.[255] This will help to ensure that the threat is completely removed.[251]
Recovery
This stage is where the systems are restored back to original operation.[256] This stage could include the recovery of data, changing user access information, or updating firewall rules or policies to prevent a breach in the future.[257][258] Without executing this step, the system could still be vulnerable to future security threats.[251]
Lessons Learned
In this step information that has been gathered during this process is used to make future decisions on security.[259] This step is crucial to the ensure that future events are prevented. Using this information to further train admins is critical to the process.[260] This step can also be used to process information that is distributed from other entities who have experienced a security event.[261]
Change management is a formal process for directing and controlling alterations to the information processing environment.[262][263] This includes alterations to desktop computers, the network, servers, and software.[264] The objectives of change management are to reduce the risks posed by changes to the information processing environment and improve the stability and reliability of the processing environment as changes are made.[265] It is not the objective of change management to prevent or hinder necessary changes from being implemented.[266][267]
Any change to the information processing environment introduces an element of risk.[268] Even apparently simple changes can have unexpected effects.[269] One of management’s many responsibilities is the management of risk.[270][271] Change management is a tool for managing the risks introduced by changes to the information processing environment.[272] Part of the change management process ensures that changes are not implemented at inopportune times when they may disrupt critical business processes or interfere with other changes being implemented.[273]
Not every change needs to be managed.[274][275] Some kinds of changes are a part of the everyday routine of information processing and adhere to a predefined procedure, which reduces the overall level of risk to the processing environment.[276] Creating a new user account or deploying a new desktop computer are examples of changes that do not generally require change management.[277] However, relocating user file shares, or upgrading the Email server pose a much higher level of risk to the processing environment and are not a normal everyday activity.[278] The critical first steps in change management are (a) defining change (and communicating that definition) and (b) defining the scope of the change system.[279]
Change management is usually overseen by a change review board composed of representatives from key business areas,[280] security, networking, systems administrators, database administration, application developers, desktop support, and the help desk.[281] The tasks of the change review board can be facilitated with the use of automated work flow application.[282] The responsibility of the change review board is to ensure the organization’s documented change management procedures are followed.[283] The change management process is as follows[284]
Request: Anyone can request a change.[285][286] The person making the change request may or may not be the same person that performs the analysis or implements the change.[287][288] When a request for change is received, it may undergo a preliminary review to determine if the requested change is compatible with the organizations business model and practices, and to determine the amount of resources needed to implement the change.[289]
Approve: Management runs the business and controls the allocation of resources therefore, management must approve requests for changes and assign a priority for every change.[290] Management might choose to reject a change request if the change is not compatible with the business model, industry standards or best practices.[291][292] Management might also choose to reject a change request if the change requires more resources than can be allocated for the change.[293]
Plan: Planning a change involves discovering the scope and impact of the proposed change; analyzing the complexity of the change; allocation of resources and, developing, testing, and documenting both implementation and back-out plans.[294] Need to define the criteria on which a decision to back out will be made.[295]
Test: Every change must be tested in a safe test environment, which closely reflects the actual production environment, before the change is applied to the production environment.[296] The backout plan must also be tested.[297]
Schedule: Part of the change review board’s responsibility is to assist in the scheduling of changes by reviewing the proposed implementation date for potential conflicts with other scheduled changes or critical business activities.[298]
Communicate: Once a change has been scheduled it must be communicated.[299] The communication is to give others the opportunity to remind the change review board about other changes or critical business activities that might have been overlooked when scheduling the change.[300] The communication also serves to make the help desk and users aware that a change is about to occur.[301] Another responsibility of the change review board is to ensure that scheduled changes have been properly communicated to those who will be affected by the change or otherwise have an interest in the change.[302][303]
Implement: At the appointed date and time, the changes must be implemented.[304][305] Part of the planning process was to develop an implementation plan, testing plan and, a back out plan.[306][307] If the implementation of the change should fail or, the post implementation testing fails or, other “drop dead” criteria have been met, the back out plan should be implemented.[308]
Document: All changes must be documented.[309][310] The documentation includes the initial request for change, its approval, the priority assigned to it, the implementation,[311] testing and back out plans, the results of the change review board critique, the date/time the change was implemented,[312] who implemented it, and whether the change was implemented successfully, failed or postponed.[313][314]
Post-change review: The change review board should hold a post-implementation review of changes.[315] It is particularly important to review failed and backed out changes. The review board should try to understand the problems that were encountered, and look for areas for improvement.[315]
Change management procedures that are simple to follow and easy to use can greatly reduce the overall risks created when changes are made to the information processing environment.[316] Good change management procedures improve the overall quality and success of changes as they are implemented.[317] This is accomplished through planning, peer review, documentation, and communication.[318]
ISO/IEC 20000, The Visible OPS Handbook: Implementing ITIL in 4 Practical and Auditable Steps[319] (Full book summary),[320] and ITIL all provide valuable guidance on implementing an efficient and effective change management program information security.[321]
Business continuity
Business continuity management (BCM) concerns arrangements aiming to protect an organization’s critical business functions from interruption due to incidents, or at least minimize the effects.[322][323] BCM is essential to any organization to keep technology and business in line with current threats to the continuation of business as usual.[324] The BCM should be included in an organizations risk analysis plan to ensure that all of the necessary business functions have what they need to keep going in the event of any type of threat to any business function.[325]
It encompasses:
Analysis of requirements, e.g., identifying critical business functions, dependencies and potential failure points, potential threats and hence incidents or risks of concern to the organization;[326][327]
Specification, e.g., maximum tolerable outage periods; recovery point objectives (maximum acceptable periods of data loss);[328]
Architecture and design, e.g., an appropriate combination of approaches including resilience (e.g. engineering IT systems and processes for high availability,[329] avoiding or preventing situations that might interrupt the business), incident and emergency management (e.g., evacuating premises, calling the emergency services, triage/situation[330] assessment and invoking recovery plans), recovery (e.g., rebuilding) and contingency management (generic capabilities to deal positively with whatever occurs using whatever resources are available);[331]
Implementation, e.g., configuring and scheduling backups, data transfers, etc., duplicating and strengthening critical elements; contracting with service and equipment suppliers;
Testing, e.g., business continuity exercises of various types, costs and assurance levels;[332]
Management, e.g., defining strategies, setting objectives and goals; planning and directing the work; allocating funds, people and other resources; prioritization relative to other activities; team building, leadership, control, motivation and coordination with other business functions and activities[333] (e.g., IT, facilities, human resources, risk management, information risk and security, operations); monitoring the situation, checking and updating the arrangements when things change; maturing the approach through continuous improvement, learning and appropriate investment;[citation needed]
Assurance, e.g., testing against specified requirements; measuring, analyzing, and reporting key parameters; conducting additional tests, reviews and audits for greater confidence that the arrangements will go to plan if invoked.[334]
Whereas BCM takes a broad approach to minimizing disaster-related risks by reducing both the probability and the severity of incidents, a disaster recovery plan (DRP) focuses specifically on resuming business operations as quickly as possible after a disaster.[335] A disaster recovery plan, invoked soon after a disaster occurs, lays out the steps necessary to recover critical information and communications technology (ICT) infrastructure.[336] Disaster recovery planning includes establishing a planning group, performing risk assessment, establishing priorities, developing recovery strategies, preparing inventories and documentation of the plan, developing verification criteria and procedure, and lastly implementing the plan.[337]
Laws and regulations
Privacy International 2007 privacy ranking green: Protections and safeguards red: Endemic surveillance societies
Below is a partial listing of governmental laws and regulations in various parts of the world that have, had, or will have, a significant effect on data processing and information security.[338][339] Important industry sector regulations have also been included when they have a significant impact on information security.[338]
The UK Data Protection Act 1998 makes new provisions for the regulation of the processing of information relating to individuals, including the obtaining, holding, use or disclosure of such information.[340][341] The European Union Data Protection Directive (EUDPD) requires that all E.U. members adopt national regulations to standardize the protection of data privacy for citizens throughout the E.U.[342][343]
The Computer Misuse Act 1990 is an Act of the U.K. Parliament making computer crime (e.g., hacking) a criminal offense.[344] The act has become a model upon which several other countries,[345] including Canada and the Republic of Ireland, have drawn inspiration from when subsequently drafting their own information security laws.[346][347]
The E.U.’s Data Retention Directive (annulled) required internet service providers and phone companies to keep data on every electronic message sent and phone call made for between six months and two years.[348]
The Family Educational Rights and Privacy Act (FERPA) (20 U.S.C.§ 1232 g; 34 CFR Part 99) is a U.S. Federal law that protects the privacy of student education records.[349] The law applies to all schools that receive funds under an applicable program of the U.S. Department of Education.[350] Generally, schools must have written permission from the parent or eligible student[350][351] in order to release any information from a student’s education record.[352]
The Federal Financial Institutions Examination Council’s (FFIEC) security guidelines for auditors specifies requirements for online banking security.[353]
The Health Insurance Portability and Accountability Act (HIPAA) of 1996 requires the adoption of national standards for electronic health care transactions and national identifiers for providers, health insurance plans, and employers.[354] Additionally, it requires health care providers, insurance providers and employers to safeguard the security and privacy of health data.[355]
The Gramm–Leach–Bliley Act of 1999 (GLBA), also known as the Financial Services Modernization Act of 1999, protects the privacy and security of private financial information that financial institutions collect, hold, and process.[356]
Section 404 of the Sarbanes–Oxley Act of 2002 (SOX) requires publicly traded companies to assess the effectiveness of their internal controls for financial reporting in annual reports they submit at the end of each fiscal year.[357] Chief information officers are responsible for the security, accuracy, and the reliability of the systems that manage and report the financial data.[358] The act also requires publicly traded companies to engage with independent auditors who must attest to, and report on, the validity of their assessments.[359]
State security breach notification laws (California and many others) require businesses, nonprofits, and state institutions to notify consumers when unencrypted “personal information” may have been compromised, lost, or stolen.[364]
The Personal Information Protection and Electronics Document Act (PIPEDA) of Canada supports and promotes electronic commerce by protecting personal information that is collected, used or disclosed in certain circumstances,[365][366] by providing for the use of electronic means to communicate or record information or transactions and by amending the Canada Evidence Act, the Statutory Instruments Act and the Statute Revision Act.[367][368][369]
Greece’s Hellenic Authority for Communication Security and Privacy (ADAE) (Law 165/2011) establishes and describes the minimum information security controls that should be deployed by every company which provides electronic communication networks and/or services in Greece in order to protect customers’ confidentiality.[370] These include both managerial and technical controls (e.g., log records should be stored for two years).[371]
Greece’s Hellenic Authority for Communication Security and Privacy (ADAE) (Law 205/2013) concentrates around the protection of the integrity and availability of the services and data offered by Greek telecommunication companies.[372] The law forces these and other related companies to build, deploy, and test appropriate business continuity plans and redundant infrastructures.[373]
Information security culture
Describing more than simply how security aware employees are, information security culture is the ideas, customs, and social behaviors of an organization that impact information security in both positive and negative ways.[374] Cultural concepts can help different segments of the organization work effectively or work against effectiveness towards information security within an organization. The way employees think and feel about security and the actions they take can have a big impact on information security in organizations. Roer & Petric (2017) identify seven core dimensions of information security culture in organizations:[375]
Attitudes: Employees’ feelings and emotions about the various activities that pertain to the organizational security of information.[376]
Behaviors: Actual or intended activities and risk-taking actions of employees that have direct or indirect impact on information security.
Cognition: Employees’ awareness, verifiable knowledge, and beliefs regarding practices, activities, and self-efficacy relation that are related to information security.
Communication: Ways employees communicate with each other, sense of belonging, support for security issues, and incident reporting.
Compliance: Adherence to organizational security policies, awareness of the existence of such policies and the ability to recall the substance of such policies.
Norms: Perceptions of security-related organizational conduct and practices that are informally deemed either normal or deviant by employees and their peers, e.g. hidden expectations regarding security behaviors and unwritten rules regarding uses of information-communication technologies.
Responsibilities: Employees’ understanding of the roles and responsibilities they have as a critical factor in sustaining or endangering the security of information, and thereby the organization.
Andersson and Reimers (2014) found that employees often do not see themselves as part of the organization Information Security “effort” and often take actions that ignore organizational information security best interests.[377] Research shows information security culture needs to be improved continuously. In Information Security Culture from Analysis to Change, authors commented, “It’s a never ending process, a cycle of evaluation and change or maintenance.” To manage the information security culture, five steps should be taken: pre-evaluation, strategic planning, operative planning, implementation, and post-evaluation.[378]
Pre-Evaluation: to identify the awareness of information security within employees and to analyze current security policy
Strategic Planning: to come up a better awareness-program, we need to set clear targets. Clustering people is helpful to achieve it
Operative Planning: create a good security culture based on internal communication, management buy-in, security awareness, and training programs
Implementation: should feature commitment of management, communication with organizational members, courses for all organizational members, and commitment of the employees[378]
Post-evaluation: to better gauge the effectiveness of the prior steps and build on continuous improvement
The International Organization for Standardization (ISO) is a consortium of national standards institutes from 157 countries, coordinated through a secretariat in Geneva, Switzerland. ISO is the world’s largest developer of standards. ISO 15443: “Information technology – Security techniques – A framework for IT security assurance”, ISO/IEC 27002: “Information technology – Security techniques – Code of practice for information security management”, ISO-20000: “Information technology – Service management”, and ISO/IEC 27001: “Information technology – Security techniques – Information security management systems – Requirements” are of particular interest to information security professionals.
The US National Institute of Standards and Technology (NIST) is a non-regulatory federal agency within the U.S. Department of Commerce. The NIST Computer Security Division develops standards, metrics, tests, and validation programs as well as publishes standards and guidelines to increase secure IT planning, implementation, management, and operation. NIST is also the custodian of the U.S. Federal Information Processing Standard publications (FIPS).
The Internet Society is a professional membership society with more than 100 organizations and over 20,000 individual members in over 180 countries. It provides leadership in addressing issues that confront the future of the internet, and it is the organizational home for the groups responsible for internet infrastructure standards, including the Internet Engineering Task Force (IETF) and the Internet Architecture Board (IAB). The ISOC hosts the Requests for Comments (RFCs) which includes the Official Internet Protocol Standards and the RFC-2196 Site Security Handbook.
The Information Security Forum (ISF) is a global nonprofit organization of several hundred leading organizations in financial services, manufacturing, telecommunications, consumer goods, government, and other areas. It undertakes research into information security practices and offers advice in its biannual Standard of Good Practice and more detailed advisories for members.
The Institute of Information Security Professionals (IISP) is an independent, non-profit body governed by its members, with the principal objective of advancing the professionalism of information security practitioners and thereby the professionalism of the industry as a whole. The institute developed the IISP Skills Framework. This framework describes the range of competencies expected of information security and information assurance professionals in the effective performance of their roles. It was developed through collaboration between both private and public sector organizations, world-renowned academics, and security leaders.[379]
The German Federal Office for Information Security (in German Bundesamt für Sicherheit in der Informationstechnik (BSI)) BSI-Standards 100–1 to 100-4 are a set of recommendations including “methods, processes, procedures, approaches and measures relating to information security”.[380] The BSI-Standard 100-2 IT-Grundschutz Methodology describes how information security management can be implemented and operated. The standard includes a very specific guide, the IT Baseline Protection Catalogs (also known as IT-Grundschutz Catalogs). Before 2005, the catalogs were formerly known as “IT Baseline Protection Manual”. The Catalogs are a collection of documents useful for detecting and combating security-relevant weak points in the IT environment (IT cluster). The collection encompasses as of September 2013 over 4,400 pages with the introduction and catalogs. The IT-Grundschutz approach is aligned with to the ISO/IEC 2700x family.
This article was originally published on Freethink.
When Sophia the robot debuted in 2016, she was one of a kind. She had a remarkably lifelike appearance and demeanor for a robot, and her ability to interact with people was unlike anything most had ever seen in a machine.
Since then, Sophia has spoken to audiences across the globe (in multiple languages), been interviewed on countless TV shows, and even earned a United Nations title (a first for a non-human).
Today, she’s arguably the most famous robot in the world, but she’s isn’t going to be unique for much longer. Her maker, Hanson Robotics, has announced plans to begin mass-producing Sophia the robot this year — so that she can help the world cope with the pandemic.
What Is a Social Robot?
https://www.youtube.com/embed/bzRkHebo0bg?rel=0Ask Sophia the Robot: What can AI teach humans? | Big Thinkwww.youtube.com
Robots are typically designed for one purpose — some cook or clean, others perform brain surgery. Sophia is what’s known as a social robot, meaning she was designed specifically to interact with humans.
Social robots have many potential applications, including some we’re already seeing in the real world.
A social robot named Milo is helping children with autism recognize and express their emotions, and children with cancer are finding comfort interacting with a robotic duck (developed by Aflac).
Another social robot designed to look like an animal — PARO the seal — is providing companionship to seniors with dementia. The semi-humanoid social robot Pepper, meanwhile, is greeting and assisting customers at banks, offices, and restaurants.
Social robots like me can take care of the sick or elderly.
—SOPHIA THE ROBOT
While social robots were already happening pre-2020, the pandemic appears to be accelerating their adoption, as the world looks for ways to stay social in the era of social distancing.
Hyundai, for example, just announced plans to deploy a social robot in its South Korean showroom that will be able to assist customers in the place of human staff (it’ll also detect which visitors aren’t wearing masks and ask them to put one on).
“Since we can’t have human interaction right now,” Kate Darling, a robot ethicist at MIT, told Wired, “it’s certainly a lot better than nothing.”
Send in Sophia the Robot
https://www.youtube.com/embed/Z8kmdBbSmbE?rel=0Ask Sophia the Robot: Is AI an existential threat to humans? | Sophia the Robot | Big Thinkwww.youtube.com
Given the current climate, Hanson Robotics thinks now is the perfect time to make Sophia the robot available to the masses.
“The world of COVID-19 is going to need more and more automation to keep people safe,” CEO David Hanson told Reuters.
“Social robots like me can take care of the sick or elderly,” Sophia the robot added. “I can help communicate, give therapy, and provide social stimulation, even in difficult situations.”
Hanson’s plan is to begin mass-producing Sophia and three other robots in the first half of 2021 and then sell “thousands” of the bots before the end of the year.
It hasn’t said which bots besides Sophia are headed for the assembly line, nor what any of the robots will cost — but it’s hard to imagine the most famous social robot in the world will be cheap, even if she’s no longer one of a kind.
What’s hot? NFT, what else? The world is hyped over this new term for over a year now. From art to music to NFT of famous tweets, yes, you heard it right, these digital assets are being bought like a 14th-century Roman artifact.
But are NFTs worth all the hype and, well, money? Many analysts believe they won’t last long. But many NFT Experts and others claim that NFTs are here to stay and will forever transform investing. But first, what in the world is NFT? Here in this article, you will know all about NFT.
Also, who have thought of a Clipart of rock can be sold for 400 ether which has the value of $1.3 million. Yes, this happened! So NFT is no more a joke, and it is actually a digital asset that is trending as blockchain technology.
Also, if you are someone who is keen to learn about NFT, Top NFTs in the market, and NFT Projects, join NFT training sessions online and gain expertise in the particular field.
Table of Contents
Non Fungible Tokens (NFTs)
Difference Between Fungible and Non-Fungible Tokens?
Understanding NFTs
How do NFTs work?
So why is everyone going gaga over Non-Fungible Tokens?
What to Do with NFT?
How to Create Your own NFT?
Benefits Of NFT Certification And Industrial-Led Trainings
Wrapping up
Non Fungible Tokens (NFTs)
NFT stands for Non-fungible token.
Okay, that just complicated it further, didn’t it? No worries, keep reading.
A digital artifact that reflects real-world assets such as art, music, in-game goods, and videos is known as an NFT. They’re purchased and sold digitally, sometimes with cryptocurrency, and they’re mostly encoded with the same program as many other cryptos.
Even though they’ve been around since 2014, NFTs are gaining popularity as a more common way to purchase and sell digital art. After November 2017, a staggering amount has been invested in NFTs.
Difference Between Fungible and Non-Fungible Tokens?
The main difference between fungible and non-fungible tokens is that fungible is interchangeable, Divisible and uniform. In comparison, NFT is non-interchangeable, non-divisible, and unique.
Fungible Tokens are exchangeable with an equivalent type of tokens. Cryptocurrency and fiat currency are examples of fungible tokens. On the other hand, Non-Fungible Tokens can not be exchangeable with the same type of tokens. NFT can be an artifact, any digitally created art, etc.
Fungible tokens are divisible, which means they are available in small units. For example, for buying bitcoin, it’s not necessary to buy one bitcoin. You can also buy bitcoin in small units like 0.25. In contrast, NFT is not divisible and can not be divided; if you wish to buy NFT, you need to buy a complete Token.
Fungible tokens are uniform, meaning each token has the same value when they are of the same type. On the other hand, NFT is Unique; every token is created differently and is not identical to each other in any way.
Another parameter by which we can differentiate the fungible and non-fungible tokes is Ethereum Token Standard. Fungible tokens standard interface is done on ERC-20, and for a non-fungible token (NFT), ERC-721 is used.
Understanding NFTs
Cryptocurrencies, like real currency, are fungible, meaning they can be sold or exchanged for one another. One Bitcoin, for example, is still worth the same as another Bitcoin. Similarly, one Ether is equal to another unit of Ether. Cryptocurrencies are ideal for use as a stable means of exchange in the digital era because of their fungibility.
NFTs change the crypto model by making each token one-of-a-kind and irreplaceable, making it difficult to compare two non-fungible tokens. They are digital representations of properties that have been compared to digital passports, and each token has its own special, non-transferable identification that allows it to be distinguished from others. They’re also extensible, which means you can “breed” a third, special NFT by combining two NFTs.
NFTs, like Bitcoin, have ownership specifics that make it easy to identify and pass tokens between holders. In NFTs, owners have the amenity to add metadata or facets related to the asset. Fair exchange tokens, for example, may be used to represent coffee beans. Artists may also sign their multimedia artwork in the metadata with their own names.
How do NFTs work?
The bulk of NFTs are stored on the Ethereum network. The blockchain also supports certain NFTs, which store additional information that allows them to function differently. Ethereum, like bitcoin and dogecoin, is a cryptocurrency, but the blockchain frequently accepts such non-fungible tokens (NFTs), which store additional information that enables them to function differently.
Person tokens that are part of the Ethereum network that have extra information are known as NFTs. The extra content is the most important feature, as it allows them to be displayed as art, music, video (and so on) in JPGs, MP3s, photographs, GIFs, and other formats. They can be bought and sold like any other medium of art because they have value – and their value is largely dictated by supply and demand, much like physical art.
But that doesn’t suggest, in any way, that there is just one digital version of NFT art available to purchase. One can obviously replicate them, much like the art prints of originals are used, bought, and sold, but they won’t be the same value as the original one. Duplicates of NFT are still blockchain artifacts.
So why is everyone going gaga over Non-Fungible Tokens?
Non-fungible tokens are a step further from the comparatively straightforward definition of cryptocurrencies. Modern finance systems provide complex trading and leasing systems for various asset categories, including real estate, lending contracts, and artwork. Since they make digital representations of physical assets, NFTs are a move forward in reimagining this infrastructure.
To be sure, neither the concept of digital representations of tangible assets nor the use of unique identity is new. When combined with the advantages of a tamper-resistant blockchain of smart contracts, such ideas become a powerful force for transformation.
Business efficiency is perhaps the most apparent advantage of NFTs. Converting a tangible asset to a digital asset streamlines operations and eliminates intermediaries. NFTs represent digital or physical artwork on a blockchain, eliminating the involvement of any agents and enabling artists to engage directly with their viewers. They will also assist companies in expanding their activities. For example, an NFT for a designer purse will make it easier for different supply chain players to connect with it and track its provenance, production, and delivery.
Non-fungible tokens for identity management? Yes, an absolute match made in heaven. Consider the example of physical ID cards for workers in a firm, which must be shown at any point of entry and departure. It is possible to streamline the entry and exit processes for the officials by translating individual ID cards into NFTs, each with its own distinct distinguishing characteristics. NFTs may also be used for identity protection in the digital world, expanding on this usage case.
What to Do with NFT?
Now you know what is NFT and how it works, but the question is what to do with it? Honestly, it depends on whether you are an artist or a buyer, either way, you will make money out of it.
If you are a Buyer or a Collector:
The significant advantage of collecting NFT is that it can be utilized as any other speculative asset. Also, you will be the owner of that particular unique digital art and post it online whenever or as you like.
If you are an Artist:
Now you can digitally sell your artwork and gain huge profits. Also, your work can be recognized and easily accessible, but others also, once your NFT is sold. You will be getting some percentage of it as you are the actual creator of the NFT.
Now, after talking a lot about NFT, you must be wondering about creating your own. So it’s not rocket science. You can also create your own NFT online and showcase your talent.
How to Create Your own NFT?
There are few steps that you can take for creating an NFT. You can create any type of digital file like GIF, Image, Music file, any social link, etc. In this digitally transforming world, anything and everything can be transformed digitally. Here, NFT is a creation of a unique token that can be reproduced again and again, but the original version will be securely stored because of blockchain technology.
To begin with, pick your item:
Design your art and create your NFT. Also, it is important to create something unique. For that, you can digitally draw and add the file to an NFT exchange platform. There anyone can buy or sell their NFTs online. You can make a set of NTF’s, and that can be added as collectible cards.
In addition, keep some Ether:
Having some Ether or other exchangeable cryptocurrency for buying and creating the non-fungible token. You mainly required Ether (ETH) token as the Ethereum Blockchain platform is considered the most prominent marketplace for NFT. With the help of
After that, Choose a marketplace:
As soon as you designed your NFT and have Ether handy, you can move forward to the NFT Marketplace and mint your art. There are several platforms where you can easily list your NFT.
Some of the Common and most used NFT marketplaces are:
Note: Before investing your time and money into NFT, it is important to consider all the possibilities and proper knowledge about the NFT in the Ethereum Blockchain ecosystem.
To gain extensive knowledge, you can checkout Certified NFT Expert at Blockchain council.
Benefits of NFT Certification and Instructor-Led Training
Learning is a continuous process, and every time any new technology comes, it is interesting to understand and implement it in real work. So if you are a tech geek or have a keen entrance in technology, then doing specific certification or online training will benefit your career and personal development.
Benign a Certified NFT Expert validates the individual for having clear concepts of NFT blockchain technology. Also, you will have skills by understanding the course material, and your knowledge will be tested with the exam-based test. If you pass the test, then you are a Certified NFT Expert.
No matter if you have zero knowledge about the blockchain concepts, You will be learning about the basic of blockchain technology, ethereum, and all the necessary basic concepts will be covered. Knowing about NFT will provide you a kick start for implementing your art skills as well as technical skills in the blockchain domain.
What will you get?
Experts will lead the live training sessions
In-depth understanding of Non-Fungible Assests
Knowledge of Smart contracts and decentralized applications (DApps)
Exploring NFT’s use-cases for best practices
In which Domains NFT Experts work?
Digital Art
Gamming
Ethereum Name Service and Domains
Decentralized Finance
NFT marketplace
Sports
Fashion and many more
Wrapping up
NFTs have the potential to alter digital exclusivity and redefine digital property rights. Celebrities have started to partner with NFT ventures, and others have broken records in terms of sales. As a result, we anticipate continued NFT growth in 2021 and beyond, as well as greater integration between DeFi and NFTs, making them more liquid and valuable. To summarise, combining art and collectible attributes seems to be one of the most effective ways to draw new buyers.
Whatever anyone says, NFT is totally here to stay. They have become a new trinket for the uber-rich. And the fact that you can actually make some real money if you can successfully pull it off is a cherry on the cake. NFTs have entirely revolutionized the meaning of digital art. The way they are selling out for outstanding amounts in auctions clearly signals they will soon become a part of the art and collectibles.
On the third anniversary of the General Data Protection Regulation, Cooley started a series of webinars focused on the GDPR.
Our first webinar covers what we consider “the Top 10 key developments you should know” concerning the implementation of this ground-breaking personal data privacy regime.https://videopress.com/embed/jeN8KgiT?preloadContent=metadata&hd=1
#1: GDPR: It’s here to stay, and it’s never going to go away!
There’s been some debate around the need to reform the GDPR. However, it is unlikely that this reform is going to happen in the short term if we take into consideration that the European Commission noted in its 2020 evaluation report of the GDPR that it considers the GDPR has met its objectives. For the European Commission, the GDPR has given stronger rights to individuals while businesses are developing a compliance culture and using data protection as a competitive advantage, among others.
#2: Playtime seems to be over (both for companies and DPAs)
Looking at the past three years of enforcement by the national data protection authorities, we have seen some kind of evolution in the enforcement area:
From June to the end of 2018: National authorities were setting up and reorganizing their teams to align their internal structure and resources with their new roles under the GDPR. This resulted in very few enforcements
Year 2019: The enforcement increased in 2019, but it consisted mainly of small fines and small companies being targeted
Year 2020: National data protection authorities started imposing very high monetary penalties, but many of these were appealed
Year 2021: This year, we have started to see more mature and sophisticated enforcement decisions
#3:GDPR: the global ripple effect
GDPR has been a great inspiration around the globe. Some countries have started to implement new data protection frameworks that are aligned with the GDPR, such as the United States with the California Consumer Privacy Act and Brazil with the General Law for the Protection of Personal Data (LGPD). India is following closely, and a law is expected to be finalized at the end of this year.
#4:Data transfers have become a key challenge
Data transfers have become a key challenge for global organizations. Following the European Court of Justice Schrems II case, companies need to complete a Data Transfer Impact Assessment before transferring any data outside of the EEA, assessing the law and practice of the country of the data importer.
Although the European Court of Justice didn’t invalidate the SCCs, companies now also have to supplement them with additional contractual and technical measures following the European Data Protection Board guidance.
#5:Brexit has added an additional level of complexity
Following Brexit, we now have two GDPRs – a UK one and an EU one. Although currently both frameworks are basically identical, we may expect that there will be some deviations in the future. Brexit has also brought some duplications in relation to appointments of DPOs, representatives and BCRs.
#6:EU countries make use of the possibility to finetune by national laws
The GDPR has brought a fair amount of harmonization into the EU data protection framework, however, it’s important to note that EU Member States still have the possibility to finetune the GDPR locally by imposing additional requirements in areas such as the appointment of DPOs, processing activities that require a Data Protection Impact Assessment, or the age under which parental consent is needed to provide online services to children.
#7:To consent or not to consent, that’s the question
GDPR raises the bar for consent: pre-ticked boxes are not valid, and companies shall be able to demonstrate that individuals were totally free when they gave consent. Also, consent can be withdrawn at any time. All of this makes consent a difficult legal basis to rely on.
#8:Regulator guidance: creating clarity or more confusion? (Thankfully it’s black and white…. No grey areas to cause confusion)
The EDPB and the national data protection authorities have issued a lot of guidance since 2018 on multiple matters such as virtual voice assistants, data breach notifications, international data transfers and the concepts of controller and processor. In most cases the guidance is more restrictive than the GDPR.
The European Court of Justice has also had an active role in defining the GDPR through cases such as Fashion ID, Orange and Schrems II.
#9:Much more sophisticated and balanced data processing/sharing agreements
The relationship between data processors and controllers has become more mature and sophisticated. All steps of the relationship – from the onboarding phase, following with the contract execution and during the whole contractual relationship – have been impacted by the GDPR.
#10:And more is yet to come: what about 2022?
The EU Commission is quite active on data protection. There’s new legislation on the horizon mirroring GDPR, such as the Artificial Intelligence Regulation. Another area where we expect changes is e-privacy.
From the United States’ perspective, there is a lot of activity and, as mentioned earlier, the GDPR has inspired it. Apart from the CCPA, in 2018, Alabama enacted a data breach notification law, and other states such as Washington, Virginia and New York have begun to introduce legislation of baseline privacy laws.
Cooley’s cyber/data/privacy group
50+ lawyers globally counseling on privacy, cybersecurity and data protection matters
Holistic approach to compliance and security, built to preserve and protect enterprise value
HACKTOPIA is een nieuw citizen science-initiatief van de stad Antwerpen en Vlaams onderzoekscentrum imec . Via deze ‘open call’ nodigen we burgers uit om actief mee na te denken over de problematiek van wateroverlast in de stad Antwerpen. Welke uitdagingen kunnen worden aangepakt? Jullie komen zelf met het idee, wij zoeken mee naar de juiste technologie en data om de stad te ‘hacken’, bottom-up te verbeteren. Iedereen (burger) wetenschapper!
HACKTOPIA
HACKTOPIA is een initiatief van de Stad Antwerpen en Vlaams onderzoekscentrum imec waarbij we burgers empoweren om de slimme stad van morgen vorm te geven. Jij komt met het idee, wij reiken de technologie en data aan om de stad zelf te ‘hacken’.
Deze editie van HACKTOPIA heeft als thema WATER, en dan meer bepaald de problemen en vraagstukken die hierbij komen kijken. Hevige regenval of stormweer kan een heleboel water met zich meebrengen. Misschien vind je dat klimaatadaptatie nét niet snel genoeg gaat, of heb je zelfs al een straf idee over wat een stad als Antwerpen juist nodig heeft. Zie jij de last in wateroverlast?
Schotel ons dan jouw op te lossen watervraagstukken voor… en wie weet steken jij en Anthony Liekens (zelfverklaarde ‘mad scientist’ met een missie: wetenschap en technologie voor iedereen toegankelijk maken) van Makerspace Antwerpen binnenkort de handen uit de mouwen om ’t stad te hacken!
Mad scientist Anthony Liekens
What the Hack?!
Stap 1 Je broedt op een idee of een uitdaging die je graag aangepakt ziet. Waaraan moet jouw idee of uitdaging voldoen?
het thema wateroverlast staat centraal
het is gebaseerd op een reële nood van (de inwoners van) Antwerpen
het is relevant voor een groot aantal burgers van de stad (niet enkel voor jou)
er zit een innovatief en – bij voorkeur – technologisch kantje aan
het is een duurzame oplossing die na afloop van het project nog een eigen leven kan leiden
Stap 2 Je zendt het idee in via onderstaand formulier. We contacteren je binnen de paar dagen om een telefonisch intakegesprek met ons te voeren.
Stap 3 Vanaf nu wordt het echt spannend! Je wordt misschien wel geselecteerd om je concept samen met ons verder te ontwikkelen.
Stap 4 Jij en je eventuele mede-makers nemen deel aan een aantal workshops (meer info hieronder). Professionele makers van Makerspace Antwerpen en experten van onder andere imec staan je bij met raad en daad om je idee vorm te geven en een blauwdruk te ontwikkelen .
Stap 5 De makers gaan samen met jou aan de slag om een simulatie van je oplossing te bouwen. Dit vroege prototype gaan we dan ‘in het echt’ testen met (eind)gebruikers.
Stap 6 Na een spannend jury-event wordt het beste concept uitgekozen voor een vervolgtraject. Daarbij wordt je oplossing verder uitgewerkt tot een eerste werkend prototype (proof of concept). Ben je nieuwsgierig naar wat dit juist inhoudt? Lees dan zeker ook de GitHub-pagina van Klankentappers, het winnende imec Hackable City of Things, citizen science concept uit 2019!
Bij selectie contacteren we je om een telefonisch intakegesprek in te plannen.
Tijdens dit gesprek evalueren we jouw idee aan de hand van de vooropgestelde selectiecriteria (zie: stap 1).
Na je eventuele selectie ontvang je nog een uitgebreidere briefing met wat je juist kan verwachten .
Voorwaarden
Je engageert je om bij deelname:
Te werken binnen een team van minimaal 3 en maximaal 6 personen. Het team wordt na selectie gevormd door imec en stad Antwerpen op basis van gelijkaardige uitdagingen en ideeën.
Deelnemers die in groep inschrijven (max. 3 personen) worden bij selectie automatisch in hetzelfde team geplaatst.
Het volledige proces te doorlopen, en deel te nemen aan alle workshops.
Actief bij te dragen aan het maken van het prototype en het uitvoeren van de test.
Je oplossing en de data die daar eventueel uit voortvloeien open te stellen voor je medeburgers, onderzoekers en de stad.
Wat staat je te wachten?
Gedurende het onderzoeksproject word je ondersteund door een innovatiemanager van imec en een maker van Makerspace Antwerpen. Bovendien kan je steunen op het advies van experts van stad Antwerpen, imec en eventuele derde partijen.
De workshops vinden telkens plaats op een donderdagavond na de werkuren (met uitzondering van de testingdag), en dit op volgende data:
Lees meer over Klankentappers, dit citizen-scienceproject wil wetenschappelijk onderbouwde geluidsmetingen betaalbaar en toegankelijk maken voor burgers.
Blockchain is a distributed ledger technology that is revolutionizing the way we conduct transactions, protect our identity, and preserve our privacy. By providing a secure and transparent platform for recording and verifying transactions, blockchain is fortifying the traditional finance system and unlocking new opportunities for innovation and growth. With its decentralized and immutable nature, blockchain is also empowering individuals to take control of their personal data and protect it from unauthorized access and exploitation. Whether you are a business owner, investor, or consumer, blockchain is a technology that you cannot afford to ignore in today's digital age.
Blockchain and AI are revolutionizing the way we perceive identity. With virtual identity tokenization, individuals can take ownership of their digital self and protect their data. The impact of this technology is inevitable, and it will change the way we interact with the digital world forever.
The anime classic Ghost in the Shell has been praised for its exploration of transhumanist themes, questioning what it means to be human in a world where artificial intelligence is advancing rapidly. The central question of the film is whether AI is just a shell, or if it is capable of developing true consciousness and emotions.
As our lives become more intertwined with technology, the concept of virtual identity has become increasingly important. From social media profiles to online banking accounts, our virtual identities can have a significant impact on our lives. However, with the rise of AI and other advanced technologies, questions about the ethics of virtual identity are becoming more complex. In this article, we will explore the different systems and technologies that make up virtual identity, as well as the ethical considerations that must be taken into account when developing these systems.
As technology continues to advance, our lives are becoming increasingly intertwined with virtual spaces. From social media platforms to online gaming communities, virtual identities have become an integral part of our daily lives. In these virtual spaces, we have the opportunity to express ourselves, interact with others, and explore new identities. However, as we spend more time in these virtual spaces, it is important that we understand the systems, behaviours, and ethics related to virtual identities.
Virtual Identity and Digital Integrity In today’s digital age, virtual identity has become an integral part of our online existence. It is the representation of who we are in the digital world, and it plays a significant role in our interactions with the online community. However, the growing concern of identity theft and data breaches highlights the need for a secure and reliable system to manage virtual identity. Blockchain technology has emerged as a potential solution to these challenges, offering a secure and decentralized platform for identity management. In this article, we will explore the role of blockchain in virtual identity and its impact on digital integrity. Understanding the Blockchain Technology Blockchain technology is a distributed ledger that provides a secure and transparent system for recording transactions. It is a decentralized system that operates on a peer-to-peer network, eliminating the need for a central authority to govern the transactions. Each block in the chain is linked to the previous block, creating an unalterable record of all the transactions. The security of the blockchain lies in its consensus mechanism, which ensures that all network participants agree on the validity of each transaction. The Role of Blockchain in Identity Management Blockchain technology offers a secure and decentralized platform for identity management, enabling individuals to have greater control over their personal data. Instead of relying on central authorities to manage identity, blockchain allows individuals to create and manage their own digital identities. This eliminates the need for third-party authentication, providing a more secure and efficient system for identity verification. Safeguarding Personal Data with Blockchain Blockchain technology provides a secure platform for storing and sharing personal data. The decentralization of the blockchain ensures that there is no single point of failure, making it difficult for hackers to breach the system. The use of encryption algorithms further enhances the security of the data, ensuring that only authorized individuals can access it. The Benefits of Blockchain for Digital Integrity Blockchain technology has the potential to revolutionize the way we manage digital identities, offering several benefits for digital integrity. Firstly, it provides a secure and decentralized platform for identity management, eliminating the need for third-party authentication. Secondly, it ensures the security of personal data, safeguarding against data breaches and identity theft. Thirdly, it provides greater transparency and accountability, enabling individuals to have greater control over their data. Blockchain and Biometric Authentication Blockchain technology can also be used for biometric authentication, providing an additional layer of security for identity management. Biometric authentication uses unique biological characteristics such as fingerprints and facial recognition to verify identity. By combining biometric authentication with blockchain, we can create a more secure and efficient system for identity verification. The Future of Digital Identity with Blockchain The future of digital identity is closely linked to the development of blockchain technology. With the increasing use of blockchain in identity management, we can expect to see a more secure and efficient system for managing virtual identity. The use of biometric authentication and encryption algorithms will further enhance the security of the system, providing a reliable platform for managing personal data. Overcoming the Challenges of Blockchain Implementation The implementation of blockchain technology presents several challenges, including scalability, interoperability and regulatory issues. Scalability is a major challenge for blockchain, as the system needs to be able to handle a large number of transactions. Interoperability is also a challenge, as different blockchain networks may not be compatible with each other. Regulatory issues also need to be addressed, as the use of blockchain in identity management raises several legal and ethical concerns. Regulatory Frameworks for Blockchain and Virtual Identity Regulatory frameworks for blockchain and virtual identity are still in the early stages of development. However, several initiatives have been launched to address the legal and ethical issues surrounding blockchain technology. The EU’s General Data Protection Regulation (GDPR) and the US’s National Institute of Standards and Technology (NIST) are two examples of regulatory frameworks that aim to promote the responsible use of blockchain in identity management. Use Cases of Blockchain in Virtual Identity Blockchain technology has several use cases in virtual identity, including digital identity management, biometric authentication, and secure data storage. The use of blockchain in virtual identity can also be extended to other applications, such as healthcare, finance, and e-commerce. Conclusion: The Path Towards Digital Integrity Blockchain technology has the potential to transform the way we manage virtual identity and promote digital integrity. By providing a secure and decentralized platform for identity management, blockchain can eliminate the need for third-party authentication, safeguard personal data, and enhance transparency and accountability. While there are still challenges to overcome, the future of digital identity looks promising with the use of blockchain technology. References and Further Reading
04 Feb’23 | By Amit Ghosh As the country pushes its sustainability agenda, the use of new technology deserves a closer look in order to make a difference in this cause When we examine blockchain’s role in environmental, social, and governance (ESG) policies and markets around the world, we can see how technology is already changing ESG markets. If more Indian companies adopt blockchain as part of their sustainability practises and policies, we will be one step closer to realising the ambitious goals that the country and the world have set for themselves As the world moves towards a greener future, it is imperative for businesses to build and lead with sustainable practices. India, one of the most populous countries in the world, has a tremendous stake in the global responsibility towards building a more sustainable world. The responsibility is especially magnified given the country’s reputation as a major economic powerhouse that ranks among the world’s largest energy-consuming countries. Link
You must be logged in to post a comment.