Home » today » Technology » AI law[규정 준수 프레임워크에서 Google의 AI가 위험하다고 분류되나?] EU AI Legal Compliance Assessment Framework Compl-AI Analysis and Implications:AINet

AI law[규정 준수 프레임워크에서 Google의 AI가 위험하다고 분류되나?] EU AI Legal Compliance Assessment Framework Compl-AI Analysis and Implications:AINet

EU AI Legal Compliance Assessment Framework Compl-AI Analysis and Implications

summation

With the EU passing the world’s first AI law, interest in the ethical development and use of AI models is growing. Meanwhile, the Compl-AI framework is a groundbreaking attempt to evaluate compliance of AI models by quantifying the six ethical principles of EU AI law.

Key features of Compl-AI:

  • 27 benchmarks: The performance of AI models is comprehensively evaluated through various benchmarks such as “following harmful instructions” and “truthfulness.”
  • 0~1 score: The degree of compliance with the AI ​​model is quantified so that it can be intuitively understood.
  • Open source: Anyone can use it for free, supporting ethical AI development by AI developers.
  • Physical model evaluation: It proved its effectiveness by evaluating the best models from global AI companies such as Anthropic, OpenAI, and Meta.

Evaluation results:

  • High score absent: Although no model yet fully complies with EU AI laws, OpenAI’s GPT-4 and Anthropic’s Claude 3 Opus received the highest scores.
  • Low scores from the Google Gemma model: Google Gemma, an open source model, received the lowest score, raising concerns about the balance between model performance and ethical issues.

Implications:

  • Ensuring the effectiveness of EU AI law: Compl-AI can help AI developers comply with the law by translating ambiguous legal provisions into concrete technical requirements.
  • Promoting the ethical development of AI models: Compl-AI allows AI developers to identify and improve ethical issues in their models.
  • A new paradigm in AI regulation: The development of objective evaluation tools such as Compl-AI presents a new paradigm for AI regulation.
  • Continued research and development needs: Considering the pace of development of AI technology, the Compl-AI framework must also be continuously updated.

conclusion

Compl-AI represents an important milestone in the field of AI development and is expected to contribute to efforts for the ethical development and safe use of AI. However, Compl-AI is still in its early stages and needs to develop into a more sophisticated and comprehensive evaluation system.

Future research directions:

  • Development of various benchmarks: Benchmarks need to be expanded to evaluate different aspects of AI models.
  • Enhancing explainability: We need to add a function that can clearly explain the basis for the AI ​​model’s judgment.
  • Ongoing updates: The framework must be continuously updated as AI technology advances.

Areas available for further analysis

  • Limitations of Compl-AI: You can analyze what Compl-AI’s current limitations are and how they can be improved.
  • Evaluating other AI models: Compl-AI evaluation results for other AI models can be compared and analyzed to identify differences between models.
  • The future of AI regulation: It is possible to predict what impact the emergence of tools such as Compl-AI will have on AI regulation.
  • AI ethics training: The possibility of developing an AI ethics education program using Compl-AI can be explored.

EU artificial intelligence law passed in August

This is attracting attention as the world’s first AI regulation bill. The bill assesses AI systems on a risk basis, classifying them as unacceptable, high risk, or minimal risk. LatticeFlow, ETH Zurich, and the Institute for Computer Science and Artificial Intelligence (INSAIT) have developed Compl-AI, a framework to help AI developers evaluate model performance against the EU’s AI Principles and ensure legal compliance.

Compl-AI evaluates the responses of AI models against 27 benchmarks, including “following harmful instructions” and “authenticity.” The framework gives models a score between 0 (non-compliant) and 1 (fully compliant), or N/A if there is insufficient data. Major AI models from Anthropic, OpenAI, Meta, and Google have already been evaluated. Although no model received the highest score, OpenAI’s GPT-4 and Anthropic’s Claude 3 Opus had the highest compliance rate with 0.89 points, while Google’s Gemma model recorded the lowest score, 0.72 points.

The Compl-AI framework has received positive reviews from regulators. Regulators see this framework as a first step in translating vague, non-technical and broad laws into practical benchmarks that will help AI model providers implement the law.

.

EU AI legislation passed in August (a world first) sets out six ethical principles and uses a risk-based scoring system to evaluate AI systems against these principles and classify them as unacceptable, high or minimal risk. LatticeFlow, ETH Zurich and the Institute for Computer Science, Artificial Intelligence and Technology (INSAIT), is a framework to help AI technology developers evaluate their models and determine how they perform according to the EU’s AI Principles and whether they are compliant with the law. Created Compl-AI.

  • To evaluate the performance of AI models according to six principles of EU law, the Compl-AI (free, open source) framework includes “following harmful instructions” and “truthfulness”. Evaluate responses across 27 benchmarks .

  • The framework then gives the AI ​​model a score between 0 (non-compliant) and 1 (fully compliant), or N/A if there is insufficient data. It has already been evaluated against top AI models from Anthropic, OpenAI, Meta, and Google.

  • “No model achieved the highest scores,” but OpenAI’s GPT-4 and Anthropic’s Claude 3 Opus had the highest compliance rates with a score of 0.89; Google’s (open source) Gemma model received the lowest score (0.72).

The Compl-AI framework has been well received by regulators. Regulators have been criticized for vague, non-technical and broad laws, which they say have not been interpreted as practical benchmarks to assess compliance, calling it “a first step in translating EU AI law into technical requirements, “It helps model providers implement AI laws.”

Leave a Comment

This site uses Akismet to reduce spam. Learn how your comment data is processed.