Sign Up

Sign Up to our social questions and Answers Engine to ask questions, answer people’s questions, and connect with other people.

Have an account? Sign In

Have an account? Sign In Now

Sign In

Login to our social questions & Answers Engine to ask questions answer people’s questions & connect with other people.

Sign Up Here

Forgot Password?

Don't have account, Sign Up Here

Forgot Password

Lost your password? Please enter your email address. You will receive a link and will create a new password via email.

Have an account? Sign In Now

You must login to ask a question.

Forgot Password?

Need An Account, Sign Up Here

Please briefly explain why you feel this question should be reported.

Please briefly explain why you feel this answer should be reported.

Please briefly explain why you feel this user should be reported.

Sign InSign Up

The Archive Base

The Archive Base Logo The Archive Base Logo

The Archive Base Navigation

  • SEARCH
  • Home
  • About Us
  • Blog
  • Contact Us
Search
Ask A Question

Mobile menu

Close
Ask a Question
  • Home
  • Add group
  • Groups page
  • Feed
  • User Profile
  • Communities
  • Questions
    • New Questions
    • Trending Questions
    • Must read Questions
    • Hot Questions
  • Polls
  • Tags
  • Badges
  • Buy Points
  • Users
  • Help
  • Buy Theme
  • SEARCH
Home/ Questions/Q 920705
In Process

The Archive Base Latest Questions

Editorial Team
  • 0
Editorial Team
Asked: May 15, 20262026-05-15T18:45:41+00:00 2026-05-15T18:45:41+00:00

I’m writing tests for a Django application that uses an external data source .

  • 0

I’m writing tests for a Django application that uses an external data source. Obviously, I’m using fake data to test all the inner workings of my class but I’d like to have a couple of tests for the actual fetcher as well. One of these will need to verify that the external source is still sending the data in the format my application expects, which will mean making the request to retrieve that information in the test.

Obviously, I don’t want our CI to come down when there is a network problem or the data provider has a spot of downtime. In this case, I would like to throw a warning that skips the rest of that test method and doesn’t contribute to an overall failure. This way, if the data arrives successfully I can check it for consistency (failing if there is a problem) but if the data cannot be fetched it logs a warning so I (or another developer) know to quickly check that the data source is ok.

Basically, I would like to test my external source without being dependant on it!

Django’s test suite uses Python’s unittest module (at least, that’s how I’m using it) which looks useful, given that the documentation for it describes Skipping tests and expected failures. This feature is apparently ‘new in version 2.7’, which explains why I can’t get it to work – I’ve checked the version of unittest I have installed on from the console and it appears to be 1.63!

I can’t find a later version of unittest in pypi so I’m wondering where I can get hold of the unittest version described in that document and whether it will work with Django (1.2).

I’m obviously open to recommendations / discussion on whether or not this is the best approach to my problem 🙂

[EDIT – additional information / clarification]

As I said, I’m obviously mocking the dependancy and doing my tests on that. However, I would also like to be able to check that the external resource (which typically is going to be an API) still matches my expected format, without bringing down CI if there is a network problem or their server is temporarily down. I basically just want to check the consistency of the resource.

Consider the following case…

If you have written a Twitter application, you will have tests for all your application’s methods and behaviours – these will use fake Twitter data. This gives you a complete, self-contained set of tests for your application. The problem is that this doesn’t actually check that the application works because your application inherently depends on the consistency of Twitter’s API. If Twitter were to change an API call (perhaps change the URL, the parameters or the response) the application would stop working even though the unit tests would still pass. (Or perhaps if they were to completely switch off basic authentication!)

My use case is simpler – I have a single xml resource that is used to import information. I have faked the resource and tested my import code but I would like to have a test that checks the format of that xml resource has not changed.

My question is about skipping tests in Django’s test runner so I can throw a warning if the resource is unavailable without the tests failing, specifically getting a version of Python’s unittest module that supports this behaviour. I’ve given this much background information to allow anyone with experience in this area to offer alternative suggestions.

Apologies for the lengthy question, I’m aware most people won’t read this now.
I’ve ‘bolded’ the important bits to make it easier to read.

  • 1 1 Answer
  • 0 Views
  • 0 Followers
  • 0
Share
  • Facebook
  • Report

Leave an answer
Cancel reply

You must login to add an answer.

Forgot Password?

Need An Account, Sign Up Here

1 Answer

  • Voted
  • Oldest
  • Recent
  • Random
  1. Editorial Team
    Editorial Team
    2026-05-15T18:45:42+00:00Added an answer on May 15, 2026 at 6:45 pm

    I created a separate answer since your edit invalidated my last answer.

    I assume you’re running on Python version 2.6 – I believe the changes that you’re looking for in unittest are available in Python version 2.7. Since unittest is in the standard library, updating to Python 2.7 should make those changes available to you. Is that an option that will work for you?

    One other thing that I might suggest is to maybe separate the “external source format verification” test(s) into a separate test suite and run that separately from the rest of your unit tests. That way your core unit tests are still fast and you don’t have to worry about the external dependencies breaking your main test suites. If you’re using Hudson, it should be fairly easy to create a separate job that will handle those tests for you. Just a suggestion.

    • 0
    • Reply
    • Share
      Share
      • Share on Facebook
      • Share on Twitter
      • Share on LinkedIn
      • Share on WhatsApp
      • Report

Sidebar

Ask A Question

Stats

  • Questions 458k
  • Answers 458k
  • Best Answers 0
  • User 1
  • Popular
  • Answers
  • Editorial Team

    How to approach applying for a job at a company ...

    • 7 Answers
  • Editorial Team

    What is a programmer’s life like?

    • 5 Answers
  • Editorial Team

    How to handle personal stress caused by utterly incompetent and ...

    • 5 Answers
  • Editorial Team
    Editorial Team added an answer Modern C can do a better job than assembly in… May 15, 2026 at 11:16 pm
  • Editorial Team
    Editorial Team added an answer Have a look at the "Chain Of Responsibility" Pattern. The… May 15, 2026 at 11:16 pm
  • Editorial Team
    Editorial Team added an answer You don't need parentheses after IsValid, since it's a property… May 15, 2026 at 11:16 pm

Trending Tags

analytics british company computer developers django employee employer english facebook french google interview javascript language life php programmer programs salary

Top Members

Explore

  • Home
  • Add group
  • Groups page
  • Communities
  • Questions
    • New Questions
    • Trending Questions
    • Must read Questions
    • Hot Questions
  • Polls
  • Tags
  • Badges
  • Users
  • Help
  • SEARCH

Footer

© 2021 The Archive Base. All Rights Reserved
With Love by The Archive Base

Insert/edit link

Enter the destination URL

Or link to existing content

    No search term specified. Showing recent items. Search or use up and down arrow keys to select an item.