How do I write JSON data stored in the dictionary data
to a file?
f = open('data.json', 'wb')
f.write(data)
This gives the error:
TypeError: must be string or buffer, not dict
19 Answers 19
data
is a Python dictionary. It needs to be encoded as JSON before writing.
Use this for maximum compatibility (Python 2 and 3):
import json
with open('data.json', 'w') as f:
json.dump(data, f)
On a modern system (i.e. Python 3 and UTF-8 support), you can write a nicer file using:
import json
with open('data.json', 'w', encoding='utf-8') as f:
json.dump(data, f, ensure_ascii=False, indent=4)
See json
documentation.
-
272@TerminalDilettante
json.dump
writes to a file or file-like object, whereasjson.dumps
returns a string.phihag– phihag08/13/2015 20:58:40Commented Aug 13, 2015 at 20:58 -
2json dump will write the json as string in the file. How can I get a json file that is not in the form of string, but just dictionary-like? Is it possible?curiouz– curiouz07/30/2023 01:57:09Commented Jul 30, 2023 at 1:57
-
1@curiouscheese Sounds like a good question to ask on StackOverflow! Make sure to explan what you mean when you say dictionary-like – maybe you can give an example file? In the end a file is just a series of bytes.phihag– phihag07/30/2023 16:07:43Commented Jul 30, 2023 at 16:07
-
4I already found the solution. What I want to be in the json file is like
[{"name":"Jack", "age":34}, {"name":"Joe", "age":54}]
, not"[{"name":"Jack", "age":34}, , {"name":"Joe", "age":54}]"
. I just used context manager withf.write(json_string)
, instead ofjson.dumps(json_string)
and it works. Because the json string that I produced is coming fromdataframe.to_json()
.curiouz– curiouz07/31/2023 03:11:11Commented Jul 31, 2023 at 3:11
To get utf8-encoded file as opposed to ascii-encoded in the accepted answer for Python 2 use:
import io, json
with io.open('data.txt', 'w', encoding='utf-8') as f:
f.write(json.dumps(data, ensure_ascii=False))
The code is simpler in Python 3:
import json
with open('data.txt', 'w') as f:
json.dump(data, f, ensure_ascii=False)
On Windows, the encoding='utf-8'
argument to open
is still necessary.
To avoid storing an encoded copy of the data in memory (result of dumps
) and to output utf8-encoded bytestrings in both Python 2 and 3, use:
import json, codecs
with open('data.txt', 'wb') as f:
json.dump(data, codecs.getwriter('utf-8')(f), ensure_ascii=False)
The codecs.getwriter
call is redundant in Python 3 but required for Python 2
Readability and size:
The use of ensure_ascii=False
gives better readability and smaller size:
>>> json.dumps({'price': '10ドル'})
'{"price": "\\u20ac10"}'
>>> json.dumps({'price': '10ドル'}, ensure_ascii=False)
'{"price": "10ドル"}'
>>> len(json.dumps({'абвгд': 1}))
37
>>> len(json.dumps({'абвгд': 1}, ensure_ascii=False).encode('utf8'))
17
Further improve readability by adding flags indent=4, sort_keys=True
(as suggested by dinos66) to arguments of dump
or dumps
. This way you'll get a nicely indented sorted structure in the json file at the cost of a slightly larger file size.
-
5The
unicode
is superfluous - the result ofjson.dumps
is already a unicode object. Note that this fails in 3.x, where this whole mess of output file mode has been cleaned up, and json always uses character strings (and character I/O) and never bytes.phihag– phihag02/14/2013 11:20:10Commented Feb 14, 2013 at 11:20
I would answer with slight modification with aforementioned answers and that is to write a prettified JSON file which human eyes can read better. For this, pass sort_keys
as True
and indent
with 4 space characters and you are good to go. Also take care of ensuring that the ascii codes will not be written in your JSON file:
with open('data.txt', 'w') as out_file:
json.dump(json_data, out_file, sort_keys = True, indent = 4,
ensure_ascii = False)
Read and write JSON files with Python 2+3; works with unicode
# -*- coding: utf-8 -*-
import json
# Make it work for Python 2+3 and with Unicode
import io
try:
to_unicode = unicode
except NameError:
to_unicode = str
# Define data
data = {'a list': [1, 42, 3.141, 1337, 'help', u'€'],
'a string': 'bla',
'another dict': {'foo': 'bar',
'key': 'value',
'the answer': 42}}
# Write JSON file
with io.open('data.json', 'w', encoding='utf8') as outfile:
str_ = json.dumps(data,
indent=4, sort_keys=True,
separators=(',', ': '), ensure_ascii=False)
outfile.write(to_unicode(str_))
# Read JSON file
with open('data.json') as data_file:
data_loaded = json.load(data_file)
print(data == data_loaded)
Explanation of the parameters of json.dump
:
indent
: Use 4 spaces to indent each entry, e.g. when a new dict is started (otherwise all will be in one line),sort_keys
: sort the keys of dictionaries. This is useful if you want to compare json files with a diff tool / put them under version control.separators
: To prevent Python from adding trailing whitespaces
With a package
Have a look at my utility package mpu
for a super simple and easy to remember one:
import mpu.io
data = mpu.io.read('example.json')
mpu.io.write('example.json', data)
Created JSON file
{
"a list":[
1,
42,
3.141,
1337,
"help",
"€"
],
"a string":"bla",
"another dict":{
"foo":"bar",
"key":"value",
"the answer":42
}
}
Common file endings
.json
Alternatives
- CSV: Super simple format (read & write)
- JSON: Nice for writing human-readable data; VERY commonly used (read & write)
- YAML: YAML is a superset of JSON, but easier to read (read & write, comparison of JSON and YAML)
- pickle: A Python serialization format (read & write)
- MessagePack (Python package): More compact representation (read & write)
- HDF5 (Python package): Nice for matrices (read & write)
- XML: exists too *sigh* (read & write)
For your application, the following might be important:
- Support by other programming languages
- Reading / writing performance
- Compactness (file size)
See also: Comparison of data serialization formats
In case you are rather looking for a way to make configuration files, you might want to read my short article Configuration files in Python
For those of you who are trying to dump greek or other "exotic" languages such as me but are also having problems (unicode errors) with weird characters such as the peace symbol (\u262E) or others which are often contained in json formated data such as Twitter's, the solution could be as follows (sort_keys is obviously optional):
import codecs, json
with codecs.open('data.json', 'w', 'utf8') as f:
f.write(json.dumps(data, sort_keys = True, ensure_ascii=False))
I don't have enough reputation to add in comments, so I just write some of my findings of this annoying TypeError here:
Basically, I think it's a bug in the json.dump()
function in Python 2 only - It can't dump a Python (dictionary / list) data containing non-ASCII characters, even you open the file with the encoding = 'utf-8'
parameter. (i.e. No matter what you do). But, json.dumps()
works on both Python 2 and 3.
To illustrate this, following up phihag's answer: the code in his answer breaks in Python 2 with exception TypeError: must be unicode, not str
, if data
contains non-ASCII characters. (Python 2.7.6, Debian):
import json
data = {u'\u0430\u0431\u0432\u0433\u0434': 1} #{u'абвгд': 1}
with open('data.txt', 'w') as outfile:
json.dump(data, outfile)
It however works fine in Python 3.
-
@AntonyHatchkins You are right for the
unicode()
part. I just realised forio
package in Python 2,write()
needsunicode
, notstr
.ibic– ibic02/12/2017 16:29:39Commented Feb 12, 2017 at 16:29 -
1This code works for me even with python2.6.6, Debian (Dec 10 2010). As well as with python2.7.9 or python3. Check it once again, plz.Antony Hatchkins– Antony Hatchkins02/21/2017 04:40:08Commented Feb 21, 2017 at 4:40
Write a data in file using JSON use json.dump() or json.dumps() used. write like this to store data in file.
import json
data = [1,2,3,4,5]
with open('no.txt', 'w') as txtfile:
json.dump(data, txtfile)
this example in list is store to a file.
-
it's similar but provide with exampleVishal Gediya– Vishal Gediya02/17/2017 07:30:23Commented Feb 17, 2017 at 7:30
json.dump(data, open('data.txt', 'wb'))
-
3This does the same thing as @phihag's answer, but is not guaranteed to work at all times. Consider such code:
f = open('1.txt', 'w'); f.write('a'); input()
. Run it and then SYGTERM it (Ctrl-Z
thenkill %1
on linux,Ctrl-Break
on Windows).1.txt
will have 0 bytes. It is because the writing was buffered and the file was neither flushed not closed at the moment when SYGTERM occurred.with
block guarantees that the file always gets closed just like 'try/finally' block does but shorter.Antony Hatchkins– Antony Hatchkins02/10/2017 10:27:07Commented Feb 10, 2017 at 10:27 -
1If you like oneliners, a cleaner option could be to use pathlib and do something like 'pathlib.Path("data.txt").write_text(json.dumps(data))'kriss– kriss11/25/2022 09:37:22Commented Nov 25, 2022 at 9:37
All previous answers are correct here is a very simple example:
#! /usr/bin/env python
import json
def write_json():
# create a dictionary
student_data = {"students":[]}
#create a list
data_holder = student_data["students"]
# just a counter
counter = 0
#loop through if you have multiple items..
while counter < 3:
data_holder.append({'id':counter})
data_holder.append({'room':counter})
counter += 1
#write the file
file_path='/tmp/student_data.json'
with open(file_path, 'w') as outfile:
print("writing file to: ",file_path)
# HERE IS WHERE THE MAGIC HAPPENS
json.dump(student_data, outfile)
outfile.close()
print("done")
write_json()
To write the JSON with indentation, "pretty print":
import json
outfile = open('data.json')
json.dump(data, outfile, indent=4)
Also, if you need to debug improperly formatted JSON, and want a helpful error message, use import simplejson
library, instead of import json
(functions should be the same)
-
Doesn't
open('data.json')
open the file in read only mode?Mihai Todor– Mihai Todor04/25/2021 22:57:05Commented Apr 25, 2021 at 22:57
Before write a dictionary into a file as a json, you have to turn that dict onto json string using json
library.
import json
data = {
"field1":{
"a": 10,
"b": 20,
},
"field2":{
"c": 30,
"d": 40,
},
}
json_data = json.dumps(json_data)
And also you can add indent to json data to look prettier.
json_data = json.dumps(json_data, indent=4)
If you want to sort keys before turning into json,
json_data = json.dumps(json_data, sort_keys=True)
You can use the combination of these two also.
Refer the json documentation here for much more features
Finally you can write into a json file
f = open('data.json', 'wb')
f.write(json_data)
For people liking oneliners (hence with
statement is not an option), a cleaner method than leaving a dangling opened file descriptor behind can be to use write_text
from pathlib
and do something like below:
pathlib.Path("data.txt").write_text(json.dumps(data))
This can be handy in some cases in contexts where statements are not allowed like:
[pathlib.Path(f"data_{x}.json").write_text(json.dumps(x)) for x in [1, 2, 3]]
I'm not claiming it should be preferred to with
(and it's likely slower), just another option.
if you are trying to write a pandas dataframe into a file using a json format i'd recommend this
destination='filepath'
saveFile = open(destination, 'w')
saveFile.write(df.to_json())
saveFile.close()
The JSON data can be written to a file as follows
hist1 = [{'val_loss': [0.5139984398465246],
'val_acc': [0.8002029867684085],
'loss': [0.593220705309384],
'acc': [0.7687131817929321]},
{'val_loss': [0.46456472964199463],
'val_acc': [0.8173602046780344],
'loss': [0.4932038113037539],
'acc': [0.8063946213802453]}]
Write to a file:
with open('text1.json', 'w') as f:
json.dump(hist1, f)
The accepted answer is fine. However, I ran into "is not json serializable" error using that.
Here's how I fixed it
with open("file-name.json", 'w')
as output:
output.write(str(response))
Although it is not a good fix as the json file it creates will not have double quotes, however it is great if you are looking for quick and dirty.
If you like prettified (indented) and keys sorted json with optional utf-8
encoding:
import json
js2 = json.dumps(j1, indent=4, sort_keys=True)
fp2 = open('16_02_json_2.json', 'w', encoding='utf-8')
fp2.write(js2)
That's it!
In case you're working on the written file, make sure to either flush or close so that all data caches are written to the disk:
fp2.flush() # you may still keep working on the file after this
fp2.close() # guaranteed data written to the disk
Below example json loading is if you don't already have the json loaded as a dictionary.
j1 = json.loads("""
{
"glossary": {
"title": "example glossary",
"GlossDiv": {
"title": "S",
"GlossList": {
"GlossEntry": {
"ID": "SGML",
"SortAs": "SGML",
"GlossTerm": "Standard Generalized Markup Language",
"Acronym": "SGML",
"Abbrev": "ISO 8879:1986",
"GlossDef": {
"para": "A meta-markup language, used to create markup languages such as DocBook.",
"GlossSeeAlso": ["GML", "XML"]
},
"GlossSee": "markup"
}
}
}
}
}
""")
Related: How to prettyprint a JSON file?
Use the json module to convert the dict to a JSON string (or write it directly):
import json
with open('data.json', 'w') as f: # text mode, not 'wb'
json.dump(data, f) # writes valid JSON
(or f.write(json.dumps(data))
if you prefer). The error happened because you tried to write a dict object instead of a string/bytes.
This is just an extra hint at the usage of json.dumps
(this is not an answer to the problem of the question, but a trick for those who have to dump numpy data types):
If there are NumPy data types in the dictionary, json.dumps()
needs an additional parameter, credits go to TypeError: Object of type 'ndarray' is not JSON serializable , and it will also fix errors like TypeError: Object of type int64 is not JSON serializable
and so on:
class NumpyEncoder(json.JSONEncoder):
""" Special json encoder for np types """
def default(self, obj):
if isinstance(obj, (np.int_, np.intc, np.intp, np.int8,
np.int16, np.int32, np.int64, np.uint8,
np.uint16, np.uint32, np.uint64)):
return int(obj)
elif isinstance(obj, (np.float_, np.float16, np.float32,
np.float64)):
return float(obj)
elif isinstance(obj, (np.ndarray,)):
return obj.tolist()
return json.JSONEncoder.default(self, obj)
And then run:
import json
#print(json.dumps(my_data[:2], indent=4, cls=NumpyEncoder)))
with open(my_dir+'/my_filename.json', 'w') as f:
json.dumps(my_data, indent=4, cls=NumpyEncoder)))
You may also want to return a string instead of a list in case of a np.array() since arrays are printed as lists that are spread over rows which will blow up the output if you have large or many arrays. The caveat: it is more difficult to access the items from the dumped dictionary later to get them back as the original array. Yet, if you do not mind having just a string of an array, this makes the dictionary more readable. Then exchange:
elif isinstance(obj, (np.ndarray,)):
return obj.tolist()
with:
elif isinstance(obj, (np.ndarray,)):
return str(obj)
or just:
else:
return str(obj)
-
What a roundabout way to do something really simpleuser32882– user3288211/02/2021 09:17:34Commented Nov 2, 2021 at 9:17
-
1@user32882 Yes, it also astonished me. Such a weak point of such a standard as json.dumps. It got downvoted perhaps because nobody expects it to be that complicated (me included), and it does not really answer the question, but in my case, I needed it.questionto42– questionto4211/02/2021 10:09:31Commented Nov 2, 2021 at 10:09
-
please take a look at the accepted answer. This shouldn't take more than a couple of lines of code.user32882– user3288211/02/2021 13:19:13Commented Nov 2, 2021 at 13:19
-
@user32882 As far as I can remember, the accepted answer cannot export numpy datatypes, which is why I added this answer. I am not sure, though, whether there is a difference regarding numpy datatypes between
json.dump
andjson.dumps
, I cannot take the time to test this now and I guess I tested this anyway. This answer shall not replace the accepted answer, but add this special case (not special at all, numpy datatypes are common).questionto42– questionto4211/02/2021 14:19:27Commented Nov 2, 2021 at 14:19 -
@user32882 Reading your comments, you have not understood this answer. The accepted answer is more or less repeated here (
dumps
instead ofdump
here so that you can use the parameters), and the class that makes numpy exports possible is just added. Nothing against downvoting for the right sake, but please think this over.questionto42– questionto4211/03/2021 14:01:56Commented Nov 3, 2021 at 14:01
IO is simpler with dummio, a package I created. Just pip install dummio
. Then
import dummio
# read
data = dummio.json.load(filepath)
# write
dummio.json.save(data, filepath=filepath)
Not that this works even if filepath is a cloud path (s3, gcs, azure). The package supports many other data types and file formats, not only dict/json.
pathlib
with one line:Path("data.json").write_text(json.dumps(data))