Good Morning inhabitants of Islandora Land,
After a long 1st of March that involved a bit of 2nd of March of re: auditing audits, pull request reviews, last minute fixes, cleaning, pulling, IRC, coffee (hot and cold) and merging, 7.x-1.9 Code Freeze is a reality!
All thanks to a wonderful group of committed committers, maintainers and volunteers!
Some conclusions (skip if more interested in the next steps):
README and LICENSE Audits happened. Green marks on the spreadsheet required some interpretation but we did it. Many Licences were normalized and even submodules and sub-sub modules that were lacking since forever some pieces got our attention.
Many enhancements and a few new things made it into this 7.x-1.9 release branch. A group of involved committers worked hard yesterday to test pulls, answer all review questions and address issues. Travis-CI played its part in making this more difficult and even requiring some small bending of rules. Which leads probably to some future discussion that maybe it could be a good idea to not make pull requests that are not part of the release or urgent the day of code freeze itself to avoid a long wait on a single queue. Just some morning ideas.
All modules (except tuque which uses a different versioning convention) got a 7.x-1.9RC1 provisional version in their .info files, to make sure nobody installs them "thinking" they are final. This is different to other releases where versioning was added later in the game. But this is a different release! Things like auditing being done before code freeze allowed us some new approaches.
Everything reported in JIRA as a Bug (and proven so) can still be Fixed and can eventually make it into our release. Wrong documentation can also be fixed, means README.md files, even when already audited, will be part of the next process and any missing info can be added to them. Again, to make any of this happen we will require two pull requests: one to 7.x and one to 7.x-1.9 branches of the affected modules. No code tasks, enhancements or improvements are permitted in 7.x-1.9 branches from now on. That is the reason we name that a freezed branch.
I will create an updated and downloadable VM in the next 10 days (deadline March 13) to allow testers to do what they are famous for: test.
Again, I know for most of you there is nothing new here, but just wanted to let you know this happened (again). And again, for those new or just curious, don't hesitate to reach out or get motivated to get involved.
Component Manager, Auditors, Committers:
Please give your petted modules a look to see if i did my job correclty. You should see a 7.x-1.9 branch in your module and .info files should have 7.x-1.9RC1 version inside. If not, I'm a human being. Ping me and I will fix that.
Thanks to all of you to make this important part of our Islandora 7.x-1.9 release happen! I have special mention to Melissa Anez which not only served as communications facilitator-translator but also as active committer doing a huge amount of pulling and merging + helping me stay sane and extending her workday to be around in case of something exploding. Also to all those who fixed their pulls to address issues yesterday, those who reviewed those pull requests and those who merged las minute pulls. You are all great! Proud to be part of this community.
Sunshine Week—a national initiative to promote a dialogue about the importance of open government and freedom of information—is on the horizon (pun most certainly intended). Held on March 12 through March 18, Sunshine Week occurs annually, coinciding with James Madison’s birthday and National Freedom of Information Day on March 16.
During Sunshine Week, hundreds of libraries, media organizations, civic groups, nonprofits, schools, federal agencies and other participants engage public discussion on the importance of open government. The purpose? To highlight the fact that government functions best when it operates transparently and that the public has the right to access information from the government.
In honor of this upcoming library holiday—and the start of women’s history month—TBT to these five women who advocated for the public’s right to know.
1. Rep. Zoe Lofgren (D-CA)
Over the course of her long career, Lofgren has consistently sponsored legislation that strengthens the public’s right to access information and her opposition to legislation that impedes First Amendment rights. Among her many activities, Rep. Lofgren is well regarded for her work on patent reform, copyright issues, digital rights and net neutrality. She successfully fought to initiate the E-rate program that provides affordable internet access for schools, libraries and rural health centers, and she is the author of legislation that would allow the unlocking of cellular phones and other digital devices to give owners more control over their devices.
2. Meredith Fuchs
As Vice President and General Counsel of the National Security Archive at George Washington University, Fuchs lead the National Security Archive though the litigation of Armstrong v. Executive Office of the President in 1989—the first White House e-mail lawsuit. The case led to numerous important precedents, including: e-mail can be defined as a record; certain contextual data associated with e-mail records must be preserved; and the public has a right to enforce the Federal Records Act. In her words: “Records of decision making at the White House, in both the federal record and presidential record components, have long been viewed as some of the most valuable records to historians and researchers.”
3. Hazel Reid O’Leary
Nominated by President Bill Clinton in 1993, O’Leary became the seventh United States Secretary of Energy and the first African American woman to serve in that office. As Secretary, O’Leary changed the department’s Office of Classification to the Office of Declassification, created an Openness Advisory Panel and encouraged the Clinton administration to end nuclear testing in the United States (nbd).
4. Eileen D. Cooke
Cooke began her career with ALA in 1964 as assistant director of the Washington Office and assumed the director position in 1972. For thirty years, she led the organization and, during her tenure, played a major role in the development, renewal and funding of key library legislation, including the Library Services and Construction Act, the Higher Education Act, the Elementary and Secondary Education Act, the Medical Library Assistance Act, the Copyright Revision Act and the establishment of the National Commission on Libraries and Information Science. Cooke was instrumental in planning ALA’s National Library Legislative Day, helping to ensure that library voices were heard loud and clear by our elected officials.
5. Patricia Glass Schuman
As the 1991-92 president-elect of ALA, Schuman launched a nationwide media campaign to focus public attention on threats to the public’s right to know—including library funding cuts, censorship and restricted access to government information—and the need to support libraries and librarians. More than 500,000 Americans called a special toll-free number or signed petitions to Congress supporting full funding for libraries. During her presidency, Schuman implemented a program of media training for ALA chapters and division leaders and founded ALA’s first Speaker’s Network.
Join us on March 15 for the 2017 James Madison Award Ceremony!
Named for President James Madison, this ALA award is presented annually near the anniversary his birth to honor individuals or groups who have championed, protected and promoted public access to government information and the public’s right to know at the national level. The 2017 James Madison Award will be presented by ALA Past President Sari Feldman at the Knight Studio at the Newseum in Washington, D.C. You can find more details about the event here.
The post Look Back, Move Forward: Five women who stood up for the public’s right to know. appeared first on District Dispatch.
Four of the 27 papers at this year's FAST conference, and both the Best Paper awards, were from UW Madison. Redundancy Does Not Imply Fault Tolerance: Analysis of Distributed Storage Reactions to Single Errors and Corruptions by Ashiwarya Ganesan et al from UW Madison wasn't a Best Paper but I would have voted for it. And kudos to the primary author for her very clear presentation. The abstract reads:
We analyze how modern distributed storage systems behave in the presence of file-system faults such as data corruption and read and write errors. We characterize eight popular distributed storage systems and uncover numerous bugs related to file-system fault tolerance. We find that modern distributed systems do not consistently use redundancy to recover from file-system faults: a single file-system fault can cause catastrophic outcomes such as data loss, corruption, and unavailability.Previous work studied the response of local file systems to media faults. At the 2005 SOSP a team from UW Madison presented IRON File Systems. They injected device faults below state-of-the-art file systems and studied the response.
commodity file system failure policies are often inconsistent, sometimes buggy, and generally inadequate in their ability to recover from partial disk failures.The experiments described in the 2017 paper were based on injecting realistic media-caused faults into the local file systems underlying the distributed storage systems in question. They inserted a FUSE-based stacked file system between the local file system and the distributed storage system being studied to inject these faults. This code is being released for others to use.
The faults they injected each affected only a single data block in the local file system of only one server in the distributed system. They did not inject faults into local file system metadata. The injected faults were corrupted data, read errors and write errors. It is notable that a single fault of this kind at a single node of the distributed storage system, which did not itself affect the replicas at other nodes, could still cause significant problems, including corrupting those replicas.
The distributed storage systems studied were Redis, ZooKeeper, Cassandra, Kafka, RethinkDB, MongoDB, LogCabin, and CockroachDB. None looked good; all failed in some ways and none effectively exploited the redundancy they maintained to recover from the failures. The paper drew five general conclusions:
- Systems employ diverse data integrity strategies, ranging from trusting the underlying file systems (RethinkDB and Redis) to using checksums (ZooKeeper, MongoDB, CockroachDB).
- Local Behavior: Faults are often undetected; even if detected, crashing is the most common local reaction
- Redundancy is underutilized: A single fault can have disastrous cluster-wide effects.
- Crash and corruption handling are entangled.
- Nuances in commonly used distributed protocols can spread corruption or data loss.
- These systems frequently use stored checksums to detect corruption, which they repair from a replica. They may therefore exhibit some of the bad behaviors of the similar processes in the distributed storage systems the paper studied.
- Systems using a distributed storage layer for long-term storage frequently place some level of trust in it, for example they may trust it to store the checksums. They do so because they believe the replication of the distributed storage system renders it fault-tolerant. This paper destroys that belief.
The paper used similar fault injection techniques to replicate and study the problem. The take-away is that file system checkers need to be transactional, a conclusion which seems obvious once it has been stated. This is especially the case since the probability of a power outage is highest immediately after power restoration. Until the checkers are transactional, their vulnerability to crashes and power failures, and the inability of the distributed storage systems above them to correctly respond poses a significant risk.
This week Frictionless Data will be at Always Already Computational: Library Collections as Data. Sierra Williams looks at the containers for museum and library collections data and why storytelling is important for understanding these collections in the digital age.
One of the great things about working at Open Knowledge International is being surrounded by fascinating projects and a community of committed individuals working at the intersection of innovative tech, social good, and underlying principles of openness in society. Working in communications, my interest and focus is on the stories we are telling about these projects. Data as subject material – whether big, small, open, or closed – does not typically come across as human-focused and if it elicits any emotive response it is usually fear of those unknown algorithms ruling our lives in unseen ways. How do we talk about data that extends beyond the fear?
One of the most powerful stories I’ve come across at the organisation is from our Frictionless Data team. They build standards, tools and best practices designed to make sharing data as easy, or frictionless, as possible. When the team discuss this work, they talk about how their Data Packages act as containers for making it easier to transport data, as shipping containerisation does for the movement of goods. Data Packages identify the metadata necessary for data to be transferred from one researcher to another, making the whole process of getting access to quality data more efficient.
Dan Fowler will be discussing Data Packages for library and museum collections data this week at the Always Already Computational: Library Collections as Data conference, designed to foster a strategic approach to developing, describing, providing access to, and encouraging reuse of museum and library collections that support computationally-driven research and teaching in areas including but not limited to Digital Humanities, Public History, Digital History, data-driven Journalism, Digital Social Science, and Digital Art History.
Librarians get the importance of structured metadata. Digital humanities scholars get that how we share our data will have lasting implications for the future. We are thrilled to see the Carnegie Museum of Art in Pittsburgh, Pennsylvania is providing its museum collections data as a downloadable Data Package and we look forward to learning more about how this data will be used.
Over the course of this year, with the continued support of a grant from the Sloan Foundation, the Frictionless Data team will be working with researchers to pilot these Data Packages and the associated tooling further. Through this process, Frictionless Data specifically hope to provide useful tools for researchers in digital humanities.
There is so much we in the open data community can learn from the history and development of museums and libraries. Some of the best storytelling I’ve read on the key role of data management in society is by Abby Smith Rumsey in her recent book When We Are No More: How Digital Memory is Shaping Our Future. She charts the growth of the Library of Alexandria and the influence it had advancing scholarship, culture and civilisation. In fact, the word museum, I learned, comes from the Temple of the Muses, or Mouseion, which housed the great Library. Rumsey also writes on the technical responsibilities and innovation of the Alexandria librarians:
“Good work requires good order. Before scholars could begin their work, the librarians had to have finished theirs. Their task was twofold: to create and maintain an intellectual organization so it would be possible to find things, and to provide physical stewardship, to keep the scrolls secure and usable over time…
Greater efficiencies in the production of scrolls over time meant the proliferation of physical objects that were valuable, fragile and once rolled up, identical in appearance…But a better solution was a technically advanced format — the codex. This is the book form we know today…”
As we look to tell stories about data, we should continue to engage with and be inspired by the work of museums and libraries, especially in how they have articulated the meaning of information processing over the years. Information has been managed for thousands of years – yielding insight from data has been and continues to be a very tangible, human issue.
Our mission at Open Knowledge International is to empower civil society organisations to use open data to improve people’s lives. Our approach to technology is not an end in itself, but rather a gateway to make a difference in the world and collaborate. Drawing inspiration from the librarians of Alexandria, we know that good work requires good order. From Data Packages to Good Tables, Frictionless Data are building tools to help build order in an otherwise chaotic information landscape.
Make sure to follow the work of the Frictionless Data team and our other projects at Open Knowledge International for more insights on how to make sense of data.
For updates from the Always Already Computational conference, follow the hashtag #aacdata on Twitter.
The Persona Task Force was appointed by LITA (Library and Information Technology Association) Board in January 2016 to better capture the varied interests, needs, and expectations of the current, future, and past LITA members. These personas were developed as a tool that will help LITA create and refine tangible benefits and actionable strategies that directly address those interests, needs, and expectations.Personas
Based on data analysis from existing data and user research (see below in the Data Collection section), the Task Force came up with 9 personas.1. New Member: Jason Creek
2. LITA Involvement Member: Rachel Stein
3. Admin-Level Member: Jennifer Torre
4. Techie Librarian: Amy Chuck
5. Librarian: Michael Zhang
6. Non member – Previous Member: Andrew Young
7. Non Member – LITA Events Participant: Steve Gonzalez
8. Non Member – Online Follower: Rebecca Miller
9. Non Member – No Connection: Carol English
Next Step: How the Personas Will Be Used
The LITA membership personas were developed to improve the overall user experience of LITA members. In February 2017, the LITA Board asked LITA committees and interest groups to select and rank three LITA member personas that are most relevant to them and to submit a concrete action suggestion(s), which each committee or interest group would like to pursue. This will allow the LITA Board to collect a good number of concrete use case scenarios for these personas and review those use cases in relation to the LITA 2016-2018 strategic plan. The LITA Board plans to select, prioritize, and implement high-impact use cases through committees and interest groups and continue to rely on the LITA personas in order to improve the overall LITA experience for its members.Timeline and Task Force Membership
The Task Force had the first virtual meeting in February and ran through Midwinter 2017. There were 11 members, including liaison to the following LITA committees:
- Assessment and Research
- Communications and Marketing
Task Force members:
- Hong Ma (Co-chair)
- Yoo Young Lee (Co-chair)
- TJ Lamanna
- Frank Skornia
- Callan Bignoli
- Lynne Edgar
- Amanda L. Goodman
- Bohyun Kim (Liaison to the LITA Board)
- Isabel Gonzalez-Smith
- Nadaleen Tempelman-Kluit
- Mark A. Beatty, Staff Support
The Task Force gathered existing data as much as possible. The data sets included sanitized LITA 2016 membership data – which contained only job title, city, state, organization type, work area, position, and membership type – sanitized LITA Forum attendee data from 2011 to 2015 – which contained only job title, city, state, organization type, and membership type – annual reports, tech-related job postings, surveys, evaluations, and work reports like 2011 LITA Emerging Leaders Project “Branding LITA: A Market Identity for the 21st Century” and the President’s Message: UX Thinking and the LITA Member Experience written by Rachel Vacek.2. User Research
The Task Force conducted 11 in-person interviews with current LITA members at the 2016 ALA Annual Conference in Orlando, FL from June 23 to June 28 for preliminary research. With help from the LITA staff, the Task Force reached out to attendees for voluntary participation in the interview. Each interview lasted between 30 to 70 minutes.
The Task Force also developed an online survey and distributed it to the following listservs from September 1 to October 3 to aim to collect opinions from current and past LITA members as well as non-members:
- LITA Blog
- LITA Social Media (Twitter and Facebook)
- LibUX Slack Channel
- Local Boston/New England Library Tech Groups
- New England Chapter of ASIS&T
- Greater Boston LibTech’rs Meetup Group
The total respondents were 477, but only 379 responses were valid to be analyzed further. Out of 379 respondents, 155 were current LITA members while 224 were either previous members or non-members. In addition, 73 volunteers were willing to participate in the follow up study.
The Task Force identified some missing areas like personal characteristics. The customized follow up short survey was sent out to 73 volunteers from November 28 to December 12 and 16 of them replied back.
ZBW German National Library of Economics: New version of multi-lingual JEL classification published in LOD
The Journal of Economic Literature Classification Scheme (JEL) was created and is maintained by the American Economic Association. The AEA provides this widely used resource freely for scholarly purposes. Thanks to André Davids (KU Leuven), who has translated the originally English-only labels of the classification to French, Spanish and German, we provide a multi-lingual version of JEL. It's lastest version (as of 2017-01) is published in the formats RDFa and RDF download files. These formats and translations are provided "as is" and are not authorized by AEA. In order to make changes in JEL tracable more easily, we have created lists of inserted and removed JEL classes in the context of the skos-history project.JEL Classification for Linked Open Dataskos-history Linked data
This Saturday 4 March marks the 7th International Open Data Day. Groups from around the world are organising events to celebrate, promote and spread the use of open data. It is a great opportunity to show the benefits of open data and encourage the adoption of open data policies in government, business and civil society. This year the focus is on four key areas: Open research data, Tracking public money flows, Open data for environment and Open data for human rights.
Open Data Day is a bottom-up initiative coordinated through a collaborative mailing list. With over 300 events currently registered on the map (a record high), it is clear that the global movement around open data is growing. Events range from meetups, talks and seminars to more hands-on demonstrations, hackathons and training sessions: anyone interested in learning more about open data is invited to participate. All outputs are open for everyone to use and re-use.
A total of 44 events received additional support through the Open Knowledge International mini-grants scheme, funded by SPARC, the Open Contracting Program of Hivos, Article 19, Hewlett Foundation and the UK Foreign & Commonwealth Office. The focus was on funding events in the four dedicated categories, as well as supporting places that will be hosting an Open Data Day event for the first time. An overview of these events is available from this blog.
It is close to impossible to summarise all the activity that will be going on this Saturday: in the list below we highlight some events happening on the different continents.
- Egypt: the Big Data, Big Questions: From Data to Knowledge in the MENA Region conference will offer a space for multidisciplinary, critical engagement, with questions formulated around knowledge and big data in the Middle East and North Africa region.
- Ghana: The Open Data Day event in Accra focuses on “Following the money: Open Data for tracking money flows on development projects in the Accra Metropolis”. Participants will dive in with data experts to gain insights into how the government spends public funds on development projects in communities in the Accra Metropolitan Area.
- Ivory Coast: Ovillage is hosting a conference about open data and a Hackathon for the creation of an open data website by and for citizens.
- Namibia: the Open & Big data Innovation Hackathon invites software developers, statisticians, tech experts and citizens to work on co-creating innovative mobile and web applications that use open and big data to enhance public service delivery, solve challenges facing communities and keep citizens informed.
- Tunesia: the workshop “Open data for Smart cities” takes place during the yearly Forum ENET’Com Feee’17
- China: Open Data Shanghai is a data-focused hackathon where participants will be given datasets and 24 hours to “hack it”. For people who are new to Data Science, there will also be educational talks by experienced data analysts. The hackathon will provide participants with a diverse range of datasets, containing a lot of data relevant to China.
- India: Datameet Hyderabad is organizing a community get together to work on generating new open data and using any existing data to make some useful tools.
- Indonesia: the Goethe-Institut, Jakarta Smart City and Web Foundation’s Open Data Lab Jakarta are organising Open is the New Smart: Making cities work for citizens, a day focused on how open data could be a key driver of the participation of citizens to jointly address urban problems more quickly and effectively.
- Iran: Transparency4Iran will host the first session in Iran on “Open Data and its’ good applications for government”.
- Malaysia: During an Open Data Exhibition in Subang Jaya, participants are going to search for data, and use the resulting data.
- Thailand: At the Open Data and Big Data Forum, hosted by the Electronic Government Agency and Social Technology Institute of Thailand, between 150-200 participants are expected.
- Australia: ODI Queensland is inviting people to an afternoon of updates & feedback on their training program and priority Open Data projects. You can also play their International Open Data Day Datopolis Game.
- French Polynesia: the Open Data Day event in Tahiti is organised by the Institut de la Statistique de la Polynésie Française and focuses on government data specifically.
- New Zealand: Open Data NZ hosts a mini-hack event at the National Library in Wellington, with a focus on environmental data.
- Estonia: Open Data Day Tartu centres around open data in science and how to use data to establish facts in the age of post-truth problematics. There will be workshops on many different tracks, from linked data methods to using data in humanities.
- Germany: Many events are taking place across Germany: a good overview is available from this Open Knowledge Germany blog. During the Heilbronn Open Data Day for example, you can work with various datasets and interfaces, such as weather data, and the organisers will provide hardware for the assembly of fine dust sensors.
- Italy: At Open Data Day Palermo, there is a full program of talks on open data applications and portals, followed by a workshop on web scraping in the afternoon.
- Spain: In Granada, the Oficina de Software Libre will organise different workshops, activities, and an open data hackathon using the data of the city and province of Granada and the open data portal of the University of Granada. Participants can work with data linking university, gender and geographical origin, air pollution data, electoral data and contracts and cultural heritage data.
- Sweden: At Brainhack Global in Stockholm, researchers from across the globe and a myriad of disciplines come together to work on innovative projects related to neuroscience and open collaboration.
- Canada: in Halifax, Nova Scotia you can participate in an Open Data Contest focused on promoting tourism, assisting new immigrants and effective management of Nova Scotia’s natural resources and protection of the environment.
- Mexico: The Failure Institute is organising the first Datatón del Fracaso (Datathon of Failure), where participants will work in teams to analyse and visualise various datasets of the institute, such as that of business closures, to better understand entrepreneurial failure.
- Mexico: Open Data Day Xalapa + Mapatón Ciudadano invites students, entrepreneurs, professionals, graphic designers, technology-based companies and anyone with an interest in open data to a full program of a Codeando Mexico workshop, talks about Social Entrepreneurship and Mobility and the presentation of the results of the first Citizen Mapatón in Xalapa.
- USA: the California Data Collaborative hosts the Open water rate spectacular event to find out more about local water prices, and whether these vary by ethnicity, income, or location. Participants will help document water prices from agencies across the state to better understand water use in California and enable regional water agencies to prepare for an uncertain future.
- USA: Code for America will hold its sixth annual “CodeAcross”: a weekend of civic hacking events hosted by over 100 cities in the Code for America network where Brigades (local communities) are located. Code for America is a national network of community organizers & technologists who are putting technology to work for the benefit of their local communities. They recently put out their Brigade Rally Cry reflecting their mission. Events are happening all over the USA, from Code for Hawaii to Code for Burlington, Vermont, but everyone interested can also hack from home or join or start another brigade (more information on this page).
- Argentina: the city of Rosario hosts their first Open Data Day this year around open government data. Different organisations have been invited to work for the advancement of data and open governments in the 21st century, working on themes such as social inclusion, open justice, sustainable mobility, responsible consumption and access to public information.
- Brazil: The Hackers Club of Teresina in conjunction with APISOL and the Open Knowledge Brazil will hold a hackathon about transparency.
- Colombia: GANA, the open government program of the Department of Nariño, hosts the NariñoVis exhibition of data visualizations that have been made using the data from the Open Data Portal of the Government of Nariño.
- Ecuador: Datalat and MediaLabUIO are organizing Open Data Day Quito, an event that includes workshops, talks and hands-on activities on opening and using open data, as well as a workshop to use local data related to the environment.
- Uruguay: the DATA Cafes community in Montevideo is organizing a meeting to renew and improve the project ATuNombre.uy, which allows you to visualise the streets of Montevideo that bear the name of a woman and to know more about those women and their stories.
- A summary on more events happening in South America is available from this blog.
More information on Open Data Day itself is available from opendataday.org. You can also follow news on the day itself through Twitter: #opendataday and #ODD17. If you are unable join an event, but would still like to participate virtually, check Why you should take 10 minutes to look at the Open Data Roadmap this Open Data Day.
Last week, OITP Director Alan Inouye and I escaped the DC swamp momentarily for a trip to NYC. We accompanied the Digital Content Working Group (DCWG) co-chairs Carolyn Anthony and Erika Linke, ALA President Julie Todaro and President-elect Jim Neal as part of ALA’s DCWG leadership contingent that meets periodically with publishers in New York City. ALA started these meetings several years ago, initially to convince trade publishers to sell e-books to public libraries. Since that time, public libraries can buy e-books, but there are still ongoing conversations with publishers about business models, contract terms and prices as well common interests and potential collaborations.
During this trip, we met with representatives from Penguin Random House and Hachette. We discussed e-book prices and audio books as well as the health of the publishing market. In general, library e-book sales have remained stable, print sales have gone up and audiobooks have skyrocketed. We also discussed hybrid business models and the reasons why publishers were not ready to offer more than one model. New aggregator platforms are coming to the library market, and some platforms do not have the capability (yet) to serve up more than one model. Of note, Hachette has recently acquired Perseus Book Group, a large mid-size trade publisher, and is working through that transition.
The ALA delegation also met with Brooklyn Public Library (BPL) leadership at their Central Library. Hosted by Director and Chief Librarian Lisa Rosenblum, our meeting also included BPL President & CEO Linda Johnson as well as two members of their Board of Trustees and senior staffers from around the Library. Our tour of the facilities and the subject-specific centers was the highlight of the trip for me. (My next blog post will describe a lot of exciting things that Brooklyn Public is doing to engage and help the community!)
During our meeting with the senior staffers, we learned more about the Brooklyn community and how the library and its 60 branches are serving them. Linda Johnson said “Our mission is literacy, in the broadest sense of the word.” One focal point, however, was the future and the improvement of the infrastructure at both the Central Library and the branches. The hallmark of renovation plans must be flexibility, given how libraries are evolving, and the path of change is far from certain. ALA provided updates on several initiatives, including the new ALA Book Club Central and our work on national public policy and lobbying. We found the meeting to be quite useful in getting a better appreciation for the priorities, needs and perspectives of a large, urban library system—essential as Washington enters into discussions about a possible infrastructure bill (given libraries’ vital role in our nation’s digital infrastructure by providing internet access) and repealing/reforming the Affordable Care Act. With major changes in the program, people will come to the library for help.
In addition, we met with the Metropolitan New York Library Council (METRO) staff, who have moved into their new building under renovation. (Grab a hard hat and keep your coat on – no heat after 4 pm.) METRO is a membership based non-profit that serves all types of libraries in the five boroughs of NYC in addition to Westchester County. The issue of space came up again as Director Nate Hill reiterated the need for flexible space. Particularly in New York City, space is at a premium, and you have to make the most out of what you have. METRO will offer new production services and programs to its members.
Another important meeting was with staff and members of the board of the Book Industry Study Group (BISG) and its new director, Brian O’Leary, a well-known publishing consultant. BISG’s mission is “to create a more informed, empowered and efficient book industry.” They are the problem solvers: through standards development or other collaborations with the book industry and libraries, they identify glitches in the supply chain from the publishers, intermediaries and libraries and then collaborate with these communities to fix problems. Fixing problems depends on “shared solutions,” so collaboration among the players is vital. With an ambitious schedule before them, ALA can expect to see more programs, research and helpful products such as BISG’s recently published BISG Quick Start Guide to Accessible Publishing. There is also a greater opportunity for ALA to partner with BISG.
So many projects and new developments, and so little time and resources! The New York City trip was a bit of a whirlwind, especially when I was trying to meet the airport shuttle at my hotel in an insanely short period of time. But I made it—back to the swamp.
The post ALA in NYC (Part 1): Working together, sharing solutions appeared first on District Dispatch.
Austin, TX DSpace community members will be attending conferences and meetings this spring where DSpace 7 UI development will be discussed. If you will be attending as well please reach out to them, "ask about DSpace 7," and find out how to get involved with community efforts around the development of DSpace 7.
New vacancy listings are posted weekly on Wednesday at approximately 12 noon Central Time. They appear under New This Week and under the appropriate regional listing. Postings remain on the LITA Job Site for a minimum of four weeks.
New This Week
Visit the LITA Job Site for more available jobs and for information on submitting a job posting.
This October it will be 15 years since I wrote the oft-cited “MARC Must Die” column for Library Journal. At the time I took a lot of heat for that sentiment (and speaking engagements to defend it) but over time the profession has pretty much come around to realizing that it is time to do something different.
When I joined OCLC in 2007, and discovered the kind of computing power available to me as well as the single largest aggregation of library metadata in the world, I was like a kid in a candy store. What, pray tell, is actually recorded in MARC, I wondered? And now I could find out.
So in 2013 I began the “MARC Usage in WorldCat” project to reveal exactly that. I would, for particular subfields, report exactly what is contained therein, with no filtering and no normalization. This is because I felt that if we were going to move to some new way of recording and sharing our data we needed to know what we have to work with.
And often what we have to work with is not…well, standard. The same concept is often registered using many different strings which means somewhere down the line translation and normalization must happen. And of course typographical errors complicate that as well as any other automated procedure we may seek to undertake. But that’s the state of play, and only by knowing what we have to work with, in reality, will we be properly armed to make the right decisions.
It has been interesting to see changes over time as well. Some fields are new, and their use has climbed over the years. Others have been deprecated and have declined in use. Once I even detected a subfield that had gone into decline after being made obsolete and then oddly shot back up again. My colleagues traced the problem to one particular record contributor and asked them to desist in using it.
As we move to a linked data world we will need to do a lot of computational maneuvering to create links where none previously existed. This work is complicated by a diverse, non-standard, and error-riddled past. That’s why we need to know exactly what we are dealing with, and this project aims to make it possible to move into the future from a position of knowledge and understanding. As always, if you would like to see a report on a particular subfield just let me know.About Roy Tennant
Roy Tennant works on projects related to improving the technological infrastructure of libraries, museums, and archives.Mail | Web | Twitter | Facebook | LinkedIn | Flickr | YouTube | More Posts (98)
I’m excited to announce this week is my first week working for the Othmer Library division at the Chemical Heritage Foundation. CHF’s name isn’t necessarily optimal at explaining what the organization does: It’s actually an independent history of science institute (not in fact focusing exclusively on chemistry), with a museum, significant archival collection, and ongoing research activities. As they say on their home page, “The Chemical Heritage Foundation is dedicated to explaining a simple truth: science has a past and our future depends on it.” That’s a nice way to put it.
I’ll be working, at least initially, mostly on the Hydra/Sufia stack. CHF has been a significant contributor to this open source staff already (with existing developer Anna Headley, who’s still here at CHF, fortunately for us all!), and I am happy to be at a place that prioritizes open source contributions. CHF has some really interesting collections (medieval alchemical manuscripts? Oral histories from scientists? that and more), which aren’t available on the web yet — but we’re working on it.
CHF is located in Philadelphia, but I’ll still be in Baltimore, mostly working remotely, with occasional but regular visits. (Conveniently, Philadelphia is only about 100 miles from Baltimore).
And I’m very happy to be back in the library community. It’s a little bit less confusing now if I tell people I’m “a librarian”. Just a little. I definitely missed being in the library world, and the camaraderie and collaboration of the library open source tech community in my year+ I was mostly absent from it — it really is something special.
I have nothing but good things to say about Friends of the Web, where I spent the past 15 months or so. I’ll miss working with my colleagues there and many aspects of the work environment. They’re really a top-notch design and Rails/React/iOS dev firm, and if you’re looking for high-quality design or app implementation work, if you need something done in either web design or app development (or both!) that you don’t have in-house resources to do, I don’t hesitate to recommend them.
Filed under: General
If you haven’t registered for the Evergreen International Conference 2017 and reserved a hotel room, do so soon. The rooms are quickly getting taken at the lovely Marriott Cincinatti Rivercenter and will run out soon!
The Digital Public Library of America is looking for excellent instructors in higher education to join its Education Advisory Committee for 2017-2018. We recently announced a new grant from the Teagle Foundation that funds the creation of a professional development curriculum and new set development for our Primary Source Sets project.
We are currently recruiting a small group of enthusiastic educators in higher education to collaborate with our current team of K-12 and higher ed instructors on this project. Specifically, we are interested in applicants working in humanities fields with US culture including English, Composition, History, Area Studies, Music, Art History, and education fields including Secondary Education, School Librarianship, and Curriculum and Instruction. Applicants can be employed in various roles in higher education including administration or as professors, lecturers, instructors, or adjuncts. Members of this group will:
- develop and deliver curriculum for professional development using DPLA and its resources for inquiry-based instruction
- collaborate with K-12 and higher ed members of the Education Advisory Committee to develop best practices for implementing the Primary Source Sets
- write new primary source sets following DPLA’s format and participate in peer review and revision (3 in 15 months)
- provide input on online professional development modules
If selected, participants are committing to:
- attend a 2-day in-person meeting on June 29-30 2017 (arriving the night of June 28 if traveling from far away) in Boston, Massachusetts
- attend six virtual meetings approximately quarterly (September 2017-November 2018)
- participate in no more than one professional development or conference presentation as scheduled
Participants will receive a $1,200 stipend (upon completion of the project in Dec 2018) for participation as well as full reimbursement for travel costs. The deadline for applications for the 2017-2018 Education Advisory Committee is March 31, 2017.
About the Digital Public Library of America
The Digital Public Library of America is a free, online library that provides access to millions of primary and secondary sources (books, photographs, audiovisual materials, maps, etc.) from libraries, archives, and museums across the United States. Since launching in April 2013, it has aggregated over 15 million items from over 2,100 institutions. The DPLA is a registered 501(c)(3) non-profit.
Open Knowledge Foundation: Announcing the Frictionless Data Tool Fund: Apply for a mini-grant to progress and improve our specifications
Today, Open Knowledge International is launching the Frictionless Data Tool Fund, a minigrant scheme offering grants of $5,000 to help extend the implementation of software libraries — code that is used to develop software programs and applications— for the Frictionless Data specifications by developing them in a range of programming languages.
The focus of the Frictionless Data project is about building tools which can help people in removing the friction removed in working with data. These library implementations will support the development of this suite. We are looking for individual developers and organizations to help us improve the specifications and implement further work. The fund will be accepting submissions from now until 31st July 2017 for work which will be completed by the end of the year.
Last year a working group was set up to progress the specifications to our first 1.0 release, and we now have an excellent foundation to add further implementations to complement our core libraries. The Tool Fund is part of the Frictionless Data project at Open Knowledge International, where we are addressing issues related to frictions that occur when working with data. We are doing this by developing a set of tools, standards, and best practices built around the Data Package standard, a containerisation format for any kind of data based on existing practices for publishing open-source software.
This is an initial list of languages we think are going to be of most benefit to the project. However, we welcome suggestions of languages which are not found here.
For some more information and to get a full understanding of what is required, take a close look at our implementation reference documentation and the v1.0 specifications of on the dedicated Tool Fund Page on the Frictionless Data site.Applications can be made from today by submitting this form.
The Frictionless Data team will select one implementer per language, and notify all applicants whether they have been successful or not at the very latest by the end of July.
The call for proposals for Access 2017 in Saskatoon is open!
Submit your proposal by April 5th. We are looking for ideas for:
- 20 min presentations (15 min presentation, ~5 min questions)
- These could be demos, theory or practice, case studies, original research, etc.
- These submissions will be double blind peer-reviewed
- 30 min panel sessions
- 5 min lightning talks
Questions? Contact us at firstname.lastname@example.org
Open Knowledge Foundation: Presenting the value of open data to NGOs: A gentle introduction to open data for Oxfam in Kenya
Open Knowledge International is a member of OD4D, a global network of leaders in the open data community, working together to develop open data solutions around the world. Here, Simeon Oriko talks about his work carrying out an embedded data fellowship with Oxfam in Kenya as part of the Open Data for Development (OD4D) programme.
For years, traditional non-governmental organisations (NGOs) in Kenya have focused on health, humanitarian and aid initiatives. Most still engage in this type of work but with an increasing shift toward adopting modern approaches that respect historical experience and consider future emerging constraints. Modern approaches such as social innovation models and using solution-based processes such as design thinking are increasingly commonplace in NGOs. Open data practices are also being adopted by some NGOs with the aim of mobilizing all their resources for transformative change.Image credit: Systems thinking about the society by Marcel Douwe Dekker CC BY 3.0
As part of an ongoing embedded data fellowship programme with Open Knowledge International, I had the opportunity late last year to meet with some Oxfam in Kenya staff and talk to them about their priorities and their organizational shift towards ‘influencing’ and ‘advocacy’. We spoke about the strategies they would like to adopt to make this shift happen. These meetings helped identify key areas of need and opportunities for the organization to adopt open data practices.
I learned from my conversations that the organization collects and uses data but most of it is not open. It was also made plain that staff faced significant challenges in their data workflow; most lack the skills or technical know-how to work with data and often heavily rely on the Monitoring and Evaluation team to handle data analysis and other key processes.
Oxfam in Kenya expressed interest in building their knowledge and capacity to work with data. One tangible outcome from my conversations was to hold workshops to sensitize them on the value of open data. We held two workshops as a result: one with Oxfam in Kenya staff and the other with some of their partner organizations.Open Data Workshop with Oxfam in Kenya Partners – January 26th, 2017
On January 13th 2017, Oxfam in Kenya hosted me for a 3-hour workshop on open data. The training focused on 2 things:
- a basic understanding of what open data is and what it does, and,
- helping them understand and map out its potential strategic value in their organization.
- Availability and Access: We spoke on the need for data to be available as a whole and at no more than a reasonable reproduction cost. We emphasized the need for data to be available in a convenient and modifiable form. It became clear that providing an easy way to download data over the internet was the preferred way of making data easily available and accessible.
- Reuse and Redistribution: The training emphasized the use of licenses or terms under which the data must be provided. Licenses and terms that permit reuse and redistribution including the intermixing with other datasets were recommended. We briefly explored using Creative Commons Licenses to license the data openly.
- Universal Participation: The idea that anyone must be able to use, re-use and redistribute the data, was emphasized as important practice in both principle and practice.
Part of the training showcased tools and visualizations to show the strategic value of open data. This was helpful in making concrete the concept of open data to the majority in the room who either had a vague idea of open data, or were learning about it for the very first time.
In follow-up meetings, there was an express interest in sharing this knowledge with some of Oxfam in Kenya’s partner organizations. This conversation led to the second workshop hosted by the Oxfam’s tax justice team for some of their partner organization: Inuka ni Sisi, Caritas Lodwar, National Taxpayers Association (NTA) and ALDEF. This training took place in Nairobi on January 26th 2017. The majority of participants for this workshop were people who ran and worked in grassroots organizations in various parts of Kenya including Wajir and Turkana.
The training followed the same format and highlighted the same themes as the one carried out for staff. We also took time during this training to help participants identify data sources and data sets that were relevant to their programs. Some participants were keen to find revenue and expenditure data from their local County Governments. A few more were interested in comparing this data with population distribution data in their regions. Most of this updated data is available on Kenya’s national Open Data portal. This granted us the opportunity to explore data publishing tools.
From conducting these trainings, I learned of a clear need to offer very simple, entry-level workshops to expose traditional NGOs and their local partner organizations to open data and its value. Many of these organizations are already working with data in one way or another but few understand and reap the benefits that data can accord them due to a lack of knowledge and capacity.
In the interest of helping these organizations mobilize their resources for transformative change, I believe actors in the global open data community should seek out local development actors and create programs and tools that easily onboard them into the open data ecosystem with a focus on building their capacity to work with data and collaborating with them to drive increase adoption of open data practices at the grassroots and community level.
In collaboration with the OD4D program, Open Knowledge International coordinates the embedded fellowship programme which places an open data expert in a CSO over 3 months to provide support in thinking through and working with open data.
Friends of mine are in the very early stages planning a sort of cooperative intended to make professional opportunities for beginning developers while getting local non-profits and small businesses in underserved communities online. Our channel in slack is #code-op – you know, co-op but with “code.”
Anyway, this idea’s been kicking around for a couple years, in part inspired by Laurie Voss’s article posted in 2012 titled “Blue-collar knowledge workers will save the economy.” It’s compelling.
The fundamental question is: if there’s 15% unemployment in one industry and 3% in another, why aren’t people switching jobs? One problem is that knowledge work requires high levels of education. A lot has been said about America’s failure to educate its children in math and sciences, and those points are all valid: a huge increase in investment in education at all stages is necessary, and a refocusing of priorities towards the sciences is a good idea.
What’s talked about less is the obvious fact that not everyone can be a manager, a programmer, a doctor or an accountant. … So that’s my first point: people aren’t switching jobs because the jobs available are too specialized and complicated for them to do.
Think about how a physical factory worked. The reason unskilled jobs in manufacturing, say, cars existed is because some very highly skilled people first got together and looked at the process of building a car and said “okay, we can automate this bit and this bit and this bit, but making machines to do this bit is way too hard”. The blue collar workers of Detroit didn’t know how to build cars: they knew how to operate one particular bit of machinery. The only people who knew how to build cars were the guys who put the machines together. …
It’s time to build knowledge factories. Where are the website factories? Obviously there are dev shops and agencies that employ hundreds of people and reap economies of scale and specialization, but those aren’t factories as I just described them. If you wanted to follow the model of a factory, then a few very skilled developers would get together and design really good, generic websites: heavily researched, carefully constructed, configured to work with minimal maintenance for long periods under high loads. Then they’d train a bunch of significantly less skilled people to do the “final assembly”: slap the images and text into place, move pages around, turn features on and off. All via highly specialized tools requiring no programming expertise whatsoever and maybe a week’s training. …
And the analogy holds true, to lesser or greater degrees, across much of the software industry. We need to stop building software for each customer and start building software assembly lines: harder, less fun, but hundreds of times more productive — and profitable. And once we’ve built the assembly lines, a new generation of blue-collar knowledge workers will be able to step up, doing the things that robots can’t do, just like they did before.
Voss’s politics bleed through but the concept of “knowledge factories” capture my fancy.
We already have to look past the daydream of web design by assembly line. In these few short years, we have already crossed the software requirement, experimenting not just with services like SquareSpace, Shopify, or Wix, that so democratize and lower the barrier of entry to websites that often look indistinguishable from many custom jobs, but we are also toying with AI in such a way that websites lay themselves out best suited for their content.
We leapfrogged “blue-collarization.”
The landscape rapidly changes. Cars were mostly built the same way for decades, but software is fluid. We can train a bunch of people to build pieces of applications, but the way applications will be built in the future will change.
It is not sufficient to know React because React won’t always be here. You need to know how to learn. You need to know what to learn.
To that end, I am not sure in our internet world “blue collar” translates. You are either a self-starter, self-learner, gig-economist – or you are not. We may be able to create jobs by training people to write code that is so modular it doesn’t require much adaptation to circumstances, but that is fleeting – an ever diminishing skill that will leave people behind.
Not to mention, what can be automated – will.
Zuboff’s Laws: 1.) Everything that can be automated will be automated. 2.) Everything that can be informated will be informated.
The reality is that although we can keep moving the bar to describe different kinds of software development, it becomes easier and easier, to the point that value and attraction — thus, the money — declines. We must wonder whether “knowledge factories,” like factories, are always on the downward slope toward automation.
If we decide what’s common to the definition of “blue collar” is repetitive work, there will be a time where there is no repetitive work – at least in software.
Instead, the question then is whether it is possible to create this same kind of “knowledge factory” supplying work that proves to be difficult to automate.
I think the answer could be in the humanities.
Federal policymaking is not exclusively about regulatory minutiae; sometimes we celebrate milestones. Yesterday Office for Information Technology Policy (OITP) Director Alan Inouye and I joined dozens of Congressional staffers, policy analysts and advocates in Cannon House Office Building in honor of the second anniversary of the Open Internet Order.
The star speaker was FCC Commissioner Mignon Clyburn, who highlighted the role of public support in 2015 that made the Order possible. Fortunately, The Hill captured verbatim the crux of Commissioner Clyburn’s comments in a quote that I (despite my furious notetaking), couldn’t quite get:
“For me it can be summed up in this way: How do we ensure that one of most inclusive, enabling, empowering platforms of our time continues to be one where our applications, products, ideas and diverse points of view have the exact same chance of being seen and heard by everyone, regardless of our class, race, economic status or where we live?”
The Open Internet Order allows the internet to remain the platform for equal access to information that it is. Everyone who uses the internet benefits from the fair playing field provided by the Order. But it holds special significance for minorities and people in low-income or underserved communities because it vests in the FCC the authority to enforce rules against anti-consumer practices that can lead to discrimination.
Imagine the internet without net neutrality. Information collected by internet service providers could be used as a proxy to enable racial profiling and target specific communities, leading to price gouging in underserved areas, predatory lending to certain groups and a host of other abuses. More importantly, a commodified internet could stifle the voices of minorities and groups who historically have struggled to be heard. Without network neutrality, this platform for a free exchange of ideas is vulnerable to a form of censorship by a handful of companies who would profit from it.To undermine network neutrality is to undercut libraries’ core values.
Hosted by four of ALA’s advocacy allies – the Center for Media Justice, Color of Change, Free Press and National Hispanic Media Coalition – the event focused on the social justice enabled by net neutrality. ALA joined these and many other organizations last December in signing the Technology Rights and Opportunity principles, “advocating for policies that ensure freedom of speech and equality of opportunity for all, while expanding the ability of the internet to drive economic opportunity and education.” These principles clearly echo the core values of libraries. To undermine network neutrality is to undercut our core values.
Actions taken and statements made by FCC Chairman Ajit Pai confirm fears about threats to network neutrality identified by Larra Clark, Krista Cox and Kara Malenfant in a District Dispatch post earlier this year. “Free speech and free expression ensured by net neutrality will not be easy to defend,” said Commissioner Clyburn, and it will take a groundswell of grassroots support to maintain the protections of the 2015 Open Internet Order. ALA will play a big role in that groundswell through strengthening ties with other advocacy groups to defend net neutrality and our core values.
The internet is now a necessity for everyone, not a luxury for a few, and as Commissioner Clyburn put it, consumers expect there to be “a cop on the beat” – somebody protecting their interests in this broadband world. The FCC, says Clyburn, should be that “cop.” Whether or not the FCC will enforce network neutrality, one thing is for certain: librarians worked hard to gain the equal access to online information ensured by the Open Internet Order, and we’re not about to give it up without a fight.