Jelou - Historial de avisos

Sistemas funcionando con normalidad

Historial de avisos

dic 2025

RCA - Latency in bots associated with an automatic component update
  • Resuelto
    Resuelto
    This incident has been resolved.
  • Investigando
    Investigando

    1. Incident Summary

    On December 31, 2025, between 6:00 a.m. and 7:00 a.m. (local time), a temporary impact on bot functionality was observed, associated with an automatic update of a system component, which caused the bots not to respond correctly during that period.

    The technical team detected the situation in a timely manner and performed a manual rollback, successfully restoring normal service operation.
    The total duration of the incident was approximately 60 minutes.

    2. Impact

    During the period between 6:00 a.m. and 7:00 a.m., the bots experienced intermittent response behavior, which may have resulted in a suboptimal experience for end users.
    No data loss or full platform unavailability was identified.

    3. Detection

    The incident was identified by the technical team through system monitoring, after detecting variations in response times following an automatic update.

    Additionally, it was identified that an automated database backup process may have contributed to the latency observed during the same period.

    4. Response

    The technical team carried out the following actions:

    • Review of recent changes applied to the environment.

    • Identification of the automatic component update as the associated factor.

    • Execution of a manual rollback to return to a stable version.

    • Continuous monitoring to confirm service stability.

    The incident was successfully resolved by the internal team.

    5. Cause

    • Automatic update of a system component, which caused a temporary impact on bot response capacity.

    • Simultaneous execution of an automated database backup process, which may have increased latency.


    • 6. Solution

    • Manual rollback of the updated component.

    • Validation of proper bot functionality.

    • Confirmation of system stability following the intervention.

Error Messages in Executions
  • Después de la muerte
    Después de la muerte

    RCA – 12/17/25

    1. Incident Summary

    On December 17, 2025, alerts were triggered related to the performance and availability of services dependent on the database. Jelou’s technical team prioritized the platform review and confirmed that certain query processes were experiencing performance degradation, impacting the proper operation of associated services.

    2. Impact

    During the incident, customers interacting with chats and integrated corporate services experienced slow responses and intermittent behavior, including increased wait times on some queries, partially affecting the overall user experience.

    3. Detection

    The incident was detected through performance monitoring systems and automated alerts, which indicated an abnormal increase in database response times. Additionally, support tickets were received from the customer support team reporting slowness and service inconsistencies.

    4. Response

    Jelou’s technical team conducted a thorough analysis of the database status, reviewing performance metrics, query load, and resource utilization. During this process, inefficiencies in query execution were identified, causing overload and negatively impacting system responsiveness.

    5. Root Cause

    The investigation determined that the incident was related to an issue in the database query optimization structure, resulting in certain operations executing inefficiently, increasing processing times and resource consumption.

    6. Resolution

    To resolve the incident, corrective adjustments were applied to the database configuration and optimization, restoring the necessary conditions for proper query processing. These actions normalized performance and restored stability to the affected services.

    7. Mitigation

    As a preventive measure, change management and validation controls for the database were reinforced, along with continuous monitoring of critical performance metrics. In addition, early warning alert mechanisms remain active to proactively detect any anomalous behavior that could impact Jelou’s services.

  • Resuelto
    Resuelto
    This incident has been resolved.
  • Investigando
    Investigando

    Some error messages are currently appearing in certain chat executions.
    Our technical team is reviewing this behavior and making the necessary adjustments to ensure proper service operation.

nov 2025

Global Cloudflare Connectivity Outage
  • Resuelto
    Resuelto
    This incident has been resolved.
  • Supervisando
    Supervisando

    Cloudfare implemented a fix and are currently monitoring the result.

  • Identificado
    Identificado

    Issue: Cloudflare is experiencing a global connectivity outage (7:10 am UTC-05).
    Impact: You may have trouble accessing apps.jelou.ai.
    Jelou Status: All Jelou services remain operational, but external access is affected.
    Temporary Solution: Please use our contingency URL to log in:
    https://apps.01lab.co/login
    Updates: Follow live updates here: https://status.jelou.ai
    Source: Cloudflare incident details: https://www.cloudflarestatus.com

oct 2025

Issue with Page Loading
  • Después de la muerte
    Después de la muerte

    RCA

    Incident Summary

    On October 30, 2025, between 15:50 and 18:39, a visual incident occurred on the platform during the deployment of the service hosted on Cloudflare.
    During that period, some static frontend assets were cached with an inconsistent version, causing visual errors in the application interface.

    The backend, APIs, and chatbots continued to operate normally, so the overall functionality and availability of the system were not affected.


    Impact

    The incident only affected the visual presentation of the platform, with no impact on service operations or user communication.
    Some users may have experienced inconsistencies in styles or interface appearance.
    The technical team immediately applied corrective actions by performing a cache purge and redeployment, restoring the normal visual display.


    Detection

    The issue was detected internally through technical team monitoring and reports of visual anomalies in the system.
    It was confirmed that the cached resource versions did not match the latest release, which caused the visual inconsistencies.


    Response

    Once the root cause was identified, the technical team performed a manual cache purge in Cloudflare and redeployed the frontend.
    This action regenerated the visual assets and restored the correct system appearance without impacting functionality.


    Root Cause

    The incident originated from an inconsistent version of static frontend assets generated during deployment, which remained temporarily cached and served to some users.
    This affected only the visual layer of the system and did not impact operations.


    Resolution and Preventive Actions

    • A cache purge was performed and the frontend was redeployed, restoring the correct visual styles.

    • An automatic integrity verification process for assets will be implemented after each deployment.

    • Frontend monitoring will be strengthened to detect cache-related visual errors earlier.

  • Resuelto
    Resuelto

    This incident has been resolved.

    We will post the RCA as an later update.

  • Supervisando
    Supervisando
    We implemented a fix and are currently monitoring the result.
  • Investigando
    Investigando

    We’re currently investigating this issue. Please avoid refreshing the page while we work on a fix.

oct 2025 a dic 2025

Siguiente