Skip to content

Feature Experimentation

Categories

JUMP TO ANOTHER FORUM

56 results found

  1. 1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    New  ·  0 comments  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  2. If a flag status is Draft and a user attempts to change the value of that flag, the UI should notify the user that they are changing a flag that is not yet in Running state.

    This will prevent cases where someone thinks they have changed a flag state, but as far as the system is concerned, they have not.

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    New  ·  0 comments  ·  UX  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  3. In the Opti ID > Usage & Billing > MAU Breakdown Dashboard, the “Experiment Name” column currently displays plain text for each experiment. While this provides identification, users must manually search for the experiment within the relevant Optimizely project to view or edit its details. This extra step slows down workflows, especially for customers or internal teams who need to investigate experiment configuration, performance, or attribution quickly.

    Solution: Enhance the “Experiment Name” field in the Opti ID MAU dashboard to display each experiment name as a clickable hyperlink. The link should point directly to the experiment’s details page. The link…

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    New  ·  0 comments  ·  UX  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  4. I know that Optimizely tries to keep bucketing consistent when traffic allocation and or distribution changes.
    It would be great to be able to see the bucketing allocation (eg. 1-5000 for A, 5001-10000 for B) in the interface.
    In this example it is simple, but when ramping up (and possibly at the same time changing the distribution of traffic) it would be great to be able to verify via the interface which buckets are being set.
    I am aware of Customer Profile Service but I see this as independent from that.

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    New  ·  0 comments  ·  UX  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  5. I know that Optimizely tries to keep bucketing consistent when traffic allocation and or distribution changes.
    It would be great to be able to see the bucketing allocation (eg. 1-5000 for A, 5001-10000 for B) in the interface.
    In this example it is simple, but when ramping up (and possibly at the same time changing the distribution of traffic) it would be great to be able to verify via the interface which buckets are being set.
    I am aware of Customer Profile Service but I see this as independent from that.

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    New  ·  0 comments  ·  UX  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  6. As of now, in Feature Experimentation, if a variant performs (very) bad, there is no way to deactivate it or set its behavior back to baseline without needing to create a new rule.
    This of course slows down experimentation speed.
    I do understand that the results of that variant are not usable after setting a variant's behavior back to baseline behavior. However, that is not the issue. The goal is to simply be able to continue letting the test run while disabling a bad-performing variant.

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    New  ·  0 comments  ·  UX  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  7. If Optimizely's CDN goes down or is inaccessible, the SDKs don't have a default fallback mechanism to evaluate feature flags without access to the datafile hosted on the CDN.

    It's possible to initialize the SDK with a cached datafile, but that requires custom logic. Ideally, Optimizely could provide a default mechanism to provide a fallback datafile (e.g., a "relay proxy" service that caches the datafile, or a default mechanism within the SDKs).

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    New  ·  0 comments  ·  SDKs  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  8. It would be great to have some kind of filters that allow a deeper dive into data. So if we wanted to see how many people bought a licence AFTER triggering a specified conversion

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    New  ·  0 comments  ·  UX  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  9. Hello!

    We very much appreciate the arrival of ratio metrics. However, they do provide limited value to us as long as Custom Event properties are not usable for either numerator or denominator.

    It would be awesome to see that added.

    Role: I am the Team Lead for our experimentation center-of-excellence and infra team.

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    New  ·  0 comments  ·  UX  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  10. Customers want to have metrics they track in their analytics in Optimizely. Today they have to manually add new event instrumentation. Would be great to integrate with other analytics and import their event tracking in Optimizely (ie. similar to Segment integration)

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    New  ·  0 comments  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  11. Filter activity notifications via webhook to only production environment.

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    New  ·  0 comments  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  12. Make SDKs compatible with OpenFeature standard.

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    New  ·  0 comments  ·  SDKs  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  13. Ability for developers to enable an event inspector to ensure instrumentation is working correctly

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    New  ·  0 comments  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  14. Be able to trigger flags if certain actions happen, or make flags dependent on other flags. Both stateless and stateful approaches.

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    New  ·  0 comments  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  15. Almost none of my customers implement long-term flags. Every time they run a new experiment, they create a new flag, and remove it after. Even if they could have reused the flag for a new experiment a month later. This creates a lot of additional work.

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    New  ·  0 comments  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  16. Help customers understand how flags/experiments are affecting site performance.

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    New  ·  1 comment  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
1 3 Next →
  • Don't see your idea?