The world's first AI laws just passed. Find out more about specifics of the historic ruling and how they might impact your team.
After three years of negotiations, the European Parliament gave its final approval on its AI Act: rules meant to create trustworthy AI systems and use cases that mitigate the safety and ethical risks posed by powerful AI models.
The Act is the world’s first formal legal framework on AI and many experts believe it could set a precedent for how many other countries approach AI legislation. The EU has almost 500 million citizens, so the legislation is expected to impact the global AI market because large AI companies will be incentivized to comply and continue offering their products in the EU. Rule enforcement for certain aspects of the Act is set to kick in this year, with most going into effect in 2025 and the final rules taking effect in 2026.
This initiative is a landmark ruling. Even still, for marketers using AI platforms like Jasper (whose interoperability means it’s likely categorized as “limited risk”), most changes revolve around transparency and copyright compliance, with minimal impact on everyday operations. However, notable requirements include disclosures for AI-driven interactions and content creation, particularly in preventing deep fakes. Jasper is committed to adhering to the new regulations, signaling a step forward in responsible AI usage. Keep reading to learn more about the EU’s AI Act.
At the heart of the Act is its risk categorization system, which regulates AI systems based on four risk levels that range from unacceptable to minimal/none. AI models and the products they power are split into these categories based on the risks they may pose to health, safety, and the fundamental rights of the EU’s citizens. Certain products may fall into multiple categories.
The AI Act addresses limited risks that may arise from general-purpose AI models. Limited risk includes large generative AI models that power chatbots and platforms. Jasper, for example, is likely defined as limited risk. These models have a wide range of applications and are increasingly foundational in EU AI systems. According to the Act, highly capable or widely used models could pose risks that may lead to accidents or misuse in cyberattacks. The spread of harmful biases across various applications could also impact many people.
The legislation emphasizes transparency and accountability for these general models around technical documentation, training data summaries, and compliance with copyright laws. Certain models will also require risk assessments, adversarial testing, and incident reporting — similar to Jasper's security protocols.
The Act features two important callouts related to transparency for applications and users leveraging these limited-risk, general-purpose AI models:
Most of the EU AI Act deals with systems deemed high-risk to EU citizens. As mentioned previously, LLMs like those used by Jasper are likely deemed “limited-risk”. This means they’re subject to smaller-scale regulations like adhering to transparency requirements and EU copyright law before releasing a new foundational model.
So in short, the Act proposes very little regulation for end-users of AI. However, there are two very important points to note:
If you’re unsure about how the Act’s reach and implications may impact your work or business, reach out to us. We also recommend contacting your legal team to get more insight.
The AI Act will officially become law after all of its member states sign off sometime this spring.
Rule enforcement will take effect in stages:
Enforcing the AI Act will primarily come down to each EU member state. States will appoint one or more national authorities to supervise AI market activities. They are encouraged to designate a national supervisory authority, which will also serve as that country’s representative in the European Artificial Intelligence Board.
An advisory forum will also provide additional technical expertise by including a diverse range of stakeholders, like representatives from the AI industry, start-ups, SMEs, civil society, and academia. Additionally, the EU Commission will create the European AI Office to oversee general-purpose AI models, collaborate with the European Artificial Intelligence Board, and benefit from guidance provided by a scientific panel of independent experts.
The Act has measures like regulatory sandboxes and provisions for real-world testing meant to help balance innovation with regulation. These initiatives intend to benefit SMEs and startups, offering them the flexibility to experiment and refine their AI systems within a controlled environment before wider deployment.
Ultimately, this legislation is a significant step forward in establishing a common framework for AI regulation. Jasper’s platform is built on the principles of security, transparency, and responsibility, and we welcome additional guidance from legislators as we continue to serve the European market. Alongside our US team, our European office is dedicated to ensuring Jasper stays compliant with all EU regulations as the AI Act is implemented.
The White House issued its own Executive Order on AI. Learn more about its major mandates and their release dates here.
High-performing marketing teams are scaling AI with structure and strategy. Learn how CMOs can lead the way to drive organization-wide impact.
July 1, 2025
|
Megan Dubin
At Cannes Lions 2025, one truth stood out: brand, trust, and authenticity are making a powerful comeback.
June 24, 2025
|
Loreal Lynch
Discover your marketing team's AI maturity. Benchmark your strategy, compare with peers, and uncover gaps with an interactive tool.
June 17, 2025
|
Megan Dubin