Let's see how to update your Drupal site between 8.x.x minor and patch versions. For example, from 8.1.2 to 8.1.3, or from 8.3.5 to 8.4.0. I hope this will help you.
If you are upgrading to Drupal version x.y.z
x -> is known as the major version number
y -> is known as the minor version number
z -> is known as the patch version number.rajamohamed Sat, 03/24/2018 - 10:31
In our last video we answered the question should I build on Drupal 7 or Drupal 8? We mentioned that we have a site currently in production on both Drupal 7 and Drupal 8. To the end-user browsing the site, it is completely transparent, they have no idea they are hitting two different Drupal instances. Today, we are going to show you how to pull this off.Read more...
Lots of people think that template engines like Twig cannot be interactively debugged. I heard this several times as an argument against template engine, and for using legacy php processing like phptemplate (standard in Drupal 7).
Well, it’s not entirely true.Tags: acquia drupal planet
With 15 years of experience in the Information Technology field, and 10 of those years focused on leadership, I’ve learned first hand the value of investing in people and setting them up for success. Before joining Mediacurrent, I started a QA department from scratch and grew it to its current size. Prior to that, I built an IT team. If you're an incoming leader, here's how to start building your own QA/IT department.
Xeno Media is pleased to announce our latest Drupal 7 contrib module, Slack to Drupal. This module imports pictures uploaded to Slack to Drupal 7 systems--thereby allowing a community of users to add content to a site while managing their daily business collaboration through the Slack app.
Zoomdata--who makes visual analytics software for big data--tasked us with coming up with a solution that allows their employees to submit images for the public website to share the company’s unique, engaging culture to aid in marketing and recruiting.
Various source platforms, including Instagram, Flickr, and Twitter, were originally considered. As we surveyed Zoomdata employees, though, we realized that Slack was the ideal source. Slack is fundamental to Zoomdata’s work culture; Its 200 employees and contractors throughout North America and Europe actively collaborating on Slack on an ongoing basis. Leveraging Slack as the source platform would allow employees to submit images in real-time without breaking their typical work/collaboration workflows and methods.
With that settled, we started researching how to integrate. Our developers researched Slack’s API and proposed two approaches: 1) Create a Slack “bot”--a virtual user that our human users could interface with. Or: 2) Integrate with a specific Slack channel. We elected the later as we could more efficiently access the files in a specific channel and Zoomdata appreciated having a single destination channel for users to come to rather than clogging other channels with off-topic bot chatter.
With the Slack-side figured out, we worked on the Drupal development. We are supporters of the Drupal Media initiative, and decided to integrate the the Drupal Media 7.x-2.0 File Entity as we do on many of our client sites. The File Entity module creates an entity like a node for each file in the system. This allows us to add fields, like Caption, Approval, Date, and Uploader. It also allows us to use, and reuse the entities in the site on other pieces of content and create views of the entities. We called this new entity Slack Image.
We also created an administration screen where an administrator can approve or disapprove images. If images are disapproved, they are removed from the system and not imported again. If approved, they are available where all the other File Entities are available.
For the Zoomdata public site, we created a view of the new Slack images that appears on their Careers page in a beautiful, modern, and responsive layout using Masonry Views, Colorbox, and GD infinite scroll plugin modules.
Our employees are always posting photos in Slack. I really wanted to share those photos with our customers, partners, prospective employees and vendors so they could get a view inside Zoomdata and know what a great team of people they’re partnering with. Jim, and the team at Xeno Media, made it possible by creating a fantastic Drupal website for us, and by developing Slack to Drupal.
Robyn Forman, Zoomdata’s VP of Digital Marketing.
Results so far have been very positive--with more than half of the company joining the channel and submissions coming from every office and department. Through Slack to Drupal, employees from throughout the organization have shown what an engaged, fun, and cutting edge culture Zoomdata really is.
We’re probably misusing the term MVP when we try to frame what we would like to see make it into core. But the actual mode of working we use there is quite an achievement. We used to grind it out endlessly, where proposed changes could be discussed endlessly, with a high risk of not committing anything at all in the end. What we’re doing now is: agree up front that it’s a good idea to improve feature X or rework interface Y. And then focus on keeping the scope as small as possible.
Yes, I, J and K are also good ideas, but we’re trying to do X here and while these are all related ideas and together would like make for a nicer whole, we should really focus on shipping X, and X alone, before turning our attention to I, J and K. If at all, because while shiny, interface Y actually presents people with more problems, so maybe we should focus on that. Though it’s never that strongly a case of either/or, and we should definately not stop iterating after the initial commit.
This is a very new and different way of working. Deliberately lowering our standards for the goal of introducing change. This is uncomfortable at times, but even that is good, because it means we’re stretching ourselves, which means we’re doing and learning new things. I’m excited and proud to see this happen. More like this.
Doing it like this means that Drupal 8.2:
- Has content moderation tools (draft! review! publish! etc.)
- Provides a new way to add new elements (blocks) to the page you’re on, without having to go to some far away corner in the admin section
- Those elements (blocks! menus! logo & site name! etc.) can then also be configured in the context of the user facing page. A side tray will show up and expose the relevant settings.
Looking forward to learn how these additions will be received and how we can improve them. In the mean time, lets add more useful and usable things to 8.3 (sample content! media handling! better dates! etc).Tags: drupaluxdrupalplanetSub title: This is a pretty radical change
The digital point of contact, the electronic business card, the online meet up for fans: A Facebook business page serves many functions. For this reason it’s important to know how to use it correctly. Here are 10 tips on how to work it.
Thinking of coming to DrupalCon Dublin this year? Why not extend your trip by a few days and stay a bit longer to take in some of the fabulous things you can go do and see in Dublin?
Here's our recommended list of things to do and see while here:1. Guinness Storehouse
The TWG coding standards committee is announcing two coding standards changes for final discussion. These appear to have reached a point close enough to consensus for final completion. The new process for proposing and ratifying changes is documented on the coding standards project page.
Official coding standards updates now ratified:
- Prefer != to <> for NOT EQUALS
- Stop disallowing camelCase for local variables / parameters
- Should we require a blank line after <?php?
Issues awaiting core approval:
- [policy] Define coding standards for anonymous functions (closures)
- Add type hinting to function declaration coding standards - sidelined on discussions around how to handle versioned coding standards (for which there is a separate issue
Issues that just need a little TLC (you can help!):
- [Policy, no patch] PHP 5.4 short array syntax coding standards - we just need some specific proposed language and this will be ratified
- [policy, no patch] Standardize indenting on chained method calls is blocked on the related coder rule issue
- [Policy, no patch] Delete permission to pad spacing in a block of related assignments needs more support - do you want this change?
These proposals will be re-evaluated during the next coding standards meeting currently scheduled for August 30th. At that point the discussion may be extended, or if clear consensus has been reached one or more policies may be dismissed or ratified and moved to the next step in the process.
For anyone who's ever looked up a definition of a Drupal term and been left wondering what it all means, here are some practical real world explanations you can use to navigate the Drupalverse. Watch this space and use comments to send us your feedback and requests.The Discipline of Dev Ops
Dev Ops, or Development Operations, is the intersection between IT managed hosting support and development. While it is a specialization in many organizations, senior developers, tech leads, and architects should be conversant in the various systems and tools to be used by your IT team or provider.
One of the primary goals of Dev Ops is to create standardized operating system iterations that are consistently reliable and easily replicable. Your unique infrastructure or hosting service plays a big role in these systems, which is why they tend to be customized to each project.
Standardized Dev Ops systems are used to create local and remote development environments, as well as staging and production environments, which all function in the same way. Having good Dev Ops systems in place means that your organization can support continuous development practices like version control and automated testing.
For any site that’s even moderately complex, having Dev Ops standards is huge. You don’t have to try to become a Dev Ops genius yourself: instead, you can find an organization like FFW to provide the level of Dev Ops help and support that is appropriate for the size and scope of your project.Defining a Display
Displays, unlike Dev Ops, are a little simpler. A Display in Drupal typically refers to how queried data is organized and shown to visitors. It is usually used in connection with a native database query referred to as a View.
One View (or database query) can have several Displays sorted in different ways. For instance, a set of queried data can be output in the following ways:
- a sortable table
- a grid
- as consecutive field sets
- in a rotating banner
- as a calendar or list of coming events
- as points on a map
… and these are only just a few examples of the many different kinds of Displays.The Details Around Distributions
A Distribution is a pre-developed assembly of database data, code, and files. Distributions commonly include saved content, configuration settings, Drupal core, contributed and custom modules, libraries, and a custom theme. It’s basically a pre-built Drupal site.
Most people first become acquainted with Distributions as different iterations of Drupal that are built for specific use cases or verticals, such as e-commerce or publishing. Many distributions are robust, production-ready applications that will save you tremendous work. They let you take advantage of the distribution sponsor’s subject matter expertise.
There are other kinds of distributions, such as ones developed mainly for marketing purposes to showcase what Drupal can do and how Drupal can be used. Both of these types of distributions have value, but it is important to differentiate between the two.
Distributions can be vetted in much the same way that a Drupal module or theme can be vetted. When evaluating a Distribution, I always like to ask the following questions:
- Who are the contributors?
- What is their experience?
- Is the project actively maintained and are new features or versions planned?
The other primary consideration when vetting a Distribution is how much complexity and effort is required to ‘unravel’ a distribution. Many organizations have found that the more fully realized distributions are difficult to customize around their specific workflows and therefore are more expensive to change than starting fresh with a more basic version of Drupal.
In recent weeks we've been making several small changes to Drupal.org: precursors to bigger things to come. First, we moved the user activity links to a user menu in the header. Next, we're moving the search function from the header to the top navigation. These changes aren't just to recover precious pixels so you can better enjoy those extra long issue summaries—these are the first step towards a new front page on Drupal.org.
As the Drupal 8 life-cycle has moved from development, to release, to adoption, we have adapted Drupal.org to support the needs of the project in the moment. And today, the need of the moment is to support the adoption journey.
As we make these changes you'll see echoes of the visual style we used when promoting the release of Drupal 8.
The Drupal wordmark region will help to define Drupal, and promote trying a demo.
A ribbon will promote contextual CTAs like learning more about Drupal 8.
The news feed will be tweaked.
DrupalCon will have a permanent home on the front page.
Community stats and featured case studies will be carried over(but may evolve).
The home page sponsorship format may change.
… a sneak preview of some new page elements and styles you'll see in the new home page.
Our first deployment will introduce the new layout and styles. Additional changes will follow as we introduce content to support our turn towards the adoption journey. Drupal evaluators beginning their adoption journey want to know who uses Drupal, and what business needs Drupal can solve. We will begin promoting specific success stories: solutions built in Drupal to meet a concrete need.What's next?
We're continuing to refine our content model and editorial workflow for the new front page. You'll see updates in the Drupal.org change notifications as we get closer to deployment.
Wondering why we're making these changes now? This turn towards the adoption journey is part of our changing priorities for the next 12 months.
Drutopia is an initiative within the Drupal project that prioritizes putting the best online tools into the hands of grassroots groups. By embracing the liberatory possibilities of free software and supporting people-centred economic models, Drutopia aims to revolutionize the way we work and cooperate.
Drutopia is at once an ethos of Drupal development and a fresh take on Drupal distributions for users to build upon, all based in a governance model that gives users a large role in the direction of the project.
Core values of the Drutopia initiative include:
- Be inclusive regarding gender, gender identity, sexual orientation, ethnicity, ability, age, religion, geography and class.
- Commit to protection of personal information and privacy and freedom from surveillance.
- Put collaboration and cooperation above competition.
- Prioritize human needs over private profit.
- Foster non-hierarchical structures and collective decision-making.
Drutopia focuses on shared solutions. Drupal excels at providing the tools to develop and distribute specialized website platforms that can be freely shared, reused, and adapted. Of the three most-used free software content management systems (CMSs) – WordPress, Joomla!, and Drupal – only Drupal has the built-in ability to package and share highly developed distributions.
Distributions are essential in attracting and meeting the needs of groups that want to support the free software movement but don’t have the technical know-how or resources to create a site from scratch. For developers, too, distributions hold a lot of potential because they do the heavy lifting of initial setup, allowing developers and site builders to bypass many hours of unnecessary effort. Drupal distributions so far have been held back by a series of factors that Drutopia aims to address.
Drutopia is about returning to Drupal’s roots in free software and progressive social change. Since its founding years, the Drupal free software project has both reflected and contributed to the democratic potential of the internet: to empower citizens to freely collaborate and organize outside the control of governments and corporate media. Long before it powered Fortune 500 sites and whitehouse.gov, Drupal was a tool of choice for small, grassroots, change-oriented groups.
This initiative aims to reclaim Drupal for the communities and groups that have always been its core users and adopters and have contributed to much of its best innovation.
Join us at drutopia.org.
The Winnipeg City’s NOW (Neighbourhoods Of Winnipeg) Portal is an initiative to create a complete neighbourhood web portal for its citizens. At the core of the project we have a set of about 47 fully linked, integrated and structured datasets of things of interests to Winnipegers. The focal point of the portal is Winnipeg’s 236 neighbourhoods, which define the main structure of the portal. The portal has six main sections: topics of interests, maps, history, census, images and economic development. The portal is meant to be used by citizens to find things of interest in their neibourhood, to learn their history, to see the images of the things of interest, to find tools to help economic development, etc.
The NOW portal is not new; Structured Dynamics was also its main technical contractor for its first release in 2013. However we just finished to help Winnipeg City’s NOW team to migrate their older NOW portal from OSF 1.x to OSF 3.x and from Drupal 6 to Drupal 7; we also trained them on the new system. Major improvements accompany this upgrade, but the user interface design is essentially the same.
The first thing I will do is to introduce each major section of the portal and I will explain the main features of each. Then I will discuss the new improvements of the portal.Datasets
A NOW portal user won’t notice any of this, but the main feature of the portal is the data it uses. The portal manages 47 datasets (and growing) of fully structured, integrated and linked datasets of things of interests to Winnipegers. What the portal does is to manage entities. Each kind of entity (swimming pools, parks, places, images, addresses, streets, etc.) are defined with multiple properties and values. Several of the entities reference other entities in other datasets (for example, an assessment parcel from the Assessment Parcels dataset references neighbourhoods entities and property addresses entities from their respective datasets).
The fact that these datasets are fully structured and integrated means that we can leverage these characteristics to create a powerful search experience by enabling filtering of the information on any of the properties, to bias the searches depending where a keyword search match occurs, etc.
Here is the list of all the 47 datasets that currently exists in the portal:
- Aboriginal Service Providers
- Neighbourhoods of Winnipeg City
- Economic Development Images
- Recreation & Leisure Images
- Neighbourhoods Images
- Volunteer Images
- Library Images
- Parks Images
- Census 2006
- Census 2001
- Winnipeg Internal Websites
- Winnipeg External Websites
- Heritage Buildings and Resources
- NOW Local Content Dataset
- Outdoor Swimming Pools
- Zoning Parcels
- School Divisions
- Property Addresses
- Wading Pools
- Electoral wards of Winnipeg City
- Assessment Parcels
- Community Centres
- Police Service Centers
- Community Gardens
- Leisure Centres
- Parks and Open Spaces
- Community Committee
- Commercial real estates
- Sports and Recreation Facilities
- Community Characterization Areas
- Indoor Swimming Pools
- Neighbourhood Clusters
- Fire and Paramedic Stations
- Bus Stops
- Fire and Paramedic Service Images
- Animal Services Images
- Skateboard Parks
- Daycare Nurseries
- Indoor Soccer Fields
- Truck Routes
- Fire Stations
- Paramedic Stations
- Spray Parks Pads
The most useful feature of the portal to me is its full-text search engine. It is simple, clean and quite effective. The search engine is configured to try to give the most relevant results a NOW portal user may be searching. For example, it will positively bias some results that comes from some specific datasets, or matches that occurs in specific property values. The goal of this biasing is to improve the quality of the returned results. This is somewhat easy to do since the context of the portal is well known and we can easily boost scoring of search results since everything is fully structured.
Another major gain is that all the search results are fully templated. The search results do not simply return a title and some description for your search results. It does template all the information the system has about the matched results, but also displays the most relevant information to the users in the search results.
For example, if I search for a indoor swimming pool, in most of the cases it may be to call the front desk to get some information about the pool. This is why different key information will be displayed directly in the search results. That way, most of the users won’t even have to click on the result to get the information they were looking for directly in the search results page.
Here is an example of a search for the keywords main street. As you can notice, you are getting different kind of results. Each result is templated to get the core information about these entities. You have the possibility to focus on particular kind of entities, or to filter by their location in specific neighbourhoods.
Now let’s see some of the kind of entities that can be searched on the portal and how they are presented to the users.
Here is an example of an assessment parcel that is located in the St. John’s neighbourhood. The address, the value, the type and the location of the parcel on a map is displayed directly into the search results.
Another kind of entity that can be searched are the property addresses. These are located on a map, the value of the parcels and the building and the zoning of the address is displayed. The property is also linked to its assessment parcel entity which can be clicked to get additional information about the parcel.
Another interesting type of entity that can be searched are the streets. What is interesting in this case is that you get the complete outline of the street directly on a map. That way you know where it starts and where it ends and where it is located in the city.
There are more than a thousand geo-localized images of all different things in the city that can be searched. A thumbnail of the image and the location of the thing that appears on the image appears in the search results.
If you were searching for a nursery for your new born child, then you can quickly see the name, location on a map and the phone number of the nursery directly in the search result.
There are just a few examples of the fifty different kind of entities that can appear like this in the search results.Mapping
The mapping tool is another powerful feature of the portal. You can search like if you were using the full-text search engine (the top search box on the portal) however you will only get the results that can be geo-localized on a map. You can also simply browse entities from a dataset or you can filter entities by their properties/values. You can persist entities you find on the map and save the map for future reference.
In the example below, it shows that someone searched for a street (main street) and then he persisted it on the map. Then he search for other things like nurseries and selected the ones that are near the street he persisted, etc. That way he can visualize the different known entities in the portal on a map to better understand where things are located in the city, what exists near a certain location, within a neighbourhood, etc.
Census information is vital to the good development of a city. They are necessary to understand the trends of a sector, who populates it, etc., such that the city and other organizations may properly plan their projects to have has much impact as possible.
These are some of the reason why one of the main section of the site is dedicated to census data. Key census indicators have been configured in the portal. Then users can select different kind of regions (neighbourhood clusters, community areas and electoral wards) to get the numbers for each of these indicators. Then they can select multiple of these regions to compare each other. A chart view and a table view is available for presenting the census data.
The City took the time to write the history of each of its neighbourhoods. In additional to that, they hired professional photographs to photograph the points of interests of the city, to geo-localize them and to write a description for each of these photos. Because of this dedication, users of the portal can learn a much about the city in general and the neighbourhood they live in. This is what the History and Image sections of the website are about.
Historic buildings are displayed on a map and they can be browsed from there.
Images of points of interests in the neighbourhood are also located on a map.
Ever wondered in which neighbourhood you live in? No problem, go on the home page, put your address in the Find your Neighbourhood section and you will know it right away. From there you can learn more about your neighbourhood like its history, the points of interest, etc.
Your address will be located on a map, and your neighbourhood will be outlined around it. Not only you will know in which neighbourhood you live, but you will also know where you live within it. From there you can click on the name of the neigbourhood to get to the neighbourhood’s page and start learning more about it like its history, to see photos of points of interest that exists in your neighbourhood, etc.
Because all the content of the portal is fully structured, it is easy to browse its content using a well defined topic structure. The city developed its own ontology that is used to help the users browse the content of the portal by browsing topics of interest. In the example below, I clicked the Economic Development node and then the Land use topic. Finally I clicked the Map button to display things that are related to land use: in this case, zoning and assessment parcels are displayed to the user.
This is another way to find meaningful and interesting content from the portal.
Depending on the topic you choose, and the kind of information related to that topic, you may end up with different options like a map, a list of links to documents related to that topic, etc.Export Content
Now that I made an overview of each of the main features of the portal, let’s go back to the geeky things. The first thing I said about this portal is that at its core, all information it manages is fully structured, integrated and linked data. If you get to the page of an entity, you have the possibility to see the underlying data that exists about it in the system. You simply have to click the Export tab at the top of the entity’s page. Then you will have access to the description of that entity in multiple different formats.
In the future, the City should (or at least I hope will) make the whole set of datasets fully downloadable. Right now you only have access to that information via that export feature per entity. I hope because this NOW portal is fully disconnected from another initiative by the city: data.winnipeg.ca, which uses Socrata. The problem is that barely any of the datasets from NOW are available on data.winnipeg.ca, and the ones that are appearing are the raw ones (semi-structured, un-documented, un-integrated and non-linked) all the normalization work, the integration work, the linkage work done by the NOW team hasn’t been leveraged to really improve the data.winnipeg.ca datasets catalog.New with the upgrades
Those who are familiar with the NOW portal will notice a few changes. The user interface did not change that much, but multiple little things got improved in the process. I will cover the most notable of these changes.
The major changes that happened are in the backend of the portal. The data management in OSF for Drupal 7 is incompatible with what was available in Drupal 6. The management of the entities became easier, the configuration of OSF networks became a breeze. A revisioning system has been added, the user interface is more intuitive, etc. There is no comparison possible. However, portal users’ won’t notice any of this, since these are all site administrator functions.
The first thing that users will notice is the completely new full-text search engine. The underlying search engine is almost the same, but the presentation is far better. All entity types have gotten their own special template, which are displayed in a special way in the search results. Most of the time results should be much more relevant, filtering is easier and cleaner. The search experience is much better in my view.
The overall site performance is much better since different caching strategies have been put in place in OSF 3.x and OSF for Drupal. This means that most of the features of the portal should react more swiftly.
Now every type of entity managed by the portal is templated: their webpage is templated in specific ways to optimize the information they want to convey to users along with their search result “mini page” when they get returned as the result of a search query.
Multi-linguality is now fully supported by the portal, however not everything is currently templated. However expect a fully translated NOW portal in French in the future.Creating a Network of Portals
One of the most interesting features that goes with this upgrade is that the NOW portal is now in a position to participate into a network of OSF instances. What does that mean? Well, it means that the NOW portal could create partnerships with other local (regional, national or international) organizations to share datasets (and their maintenance costs).
Are there other organizations that uses this kind of system? Well, there is at least another one right in Winnipeg City: MyPeg.ca, also developed by Structured Dynamics. MyPeg uses RDF to model its information and uses OSF to manage its information. MyPeg is a non-profit organization that uses census (and other indicator) data to do studies on the well being of Winnipegers. The team behind MyPeg.ca are research experts in indicator data. Their indicator datasets (which includes census data) is top notch.
Let’s hypothetize that there would be interest between the two groups to start collaborating. Let’s say that the NOW portal would like to use MyPeg’s census datasets instead of its own since they are more complete, accurate and include a larger number of important indicators. What they basically want is to outsource the creation and maintenance of the census/indicators data to a local, dedicated and highly professional organization. The only things they would need to do is to:
- Formalize their relationship by signing a usage agreement
- The NOW portal would need to configure the MyPeg.ca OSF network into their OSF for Drupal instance
- The NOW portal would need to register the datasets it want to use from MyPeg.ca.
Once these 3 steps are done, taking no more than a couple of minutes, then the system administrators of the NOW portal could start using the MyPeg.ca indicator datasets like they were existing on their own network. (The reverse could also be true for MyPeg.) Everything would be transparent to them. From then on, all the fixes and updates performed by MyPeg.ca to their indicator datasets would immediately appear on the NOW portal and accessible to its users.
This is one possibility to collaborate. Another possibility would be to simply on a routine basis (every month, every 6 months, every year) share the serialized datasets such that the NOW portal re-import the dataset from the files shared by MyPeg.ca. This is also possible since both organizations use the same Ontology to describe the indicator data. This means that no modification is required by the City to take that new information into account, they only have to import and update their local datasets. This is the beauty of ontologies.Conclusion
The new NOW portal is a great service for citizens of Winnipeg City. It is also a really good example of a web portal that leverages fully structured, integrated and linked data. To me, the NOW portal is a really good example of the features that should go along with a municipal data portal.
On August 13th, I had the pleasure of enjoying another Drupal Camp Asheville. This has become one of my favorite Drupal camps because of the location and quality of camp organization. It has the right balance of structure, while maintaining a grassroots feel that encourages open discussion and sharing.
Using the Drupal module Hook Update Deploy Tools to move node content can be an important part to a deployment strategy.
- How do I export and import nodes using Hook Update Deploy Tools? >> Read the project page or a quick how-to.
- What is the unique ID that connects an export to an import?
- What are the risks of this import export model?
- What if I am using an entity reference or a taxonomy that does not exist on production?
- Does the import show up as a revision?
- What happens if the import does not validate?
- What if the alias or path is already in use by another node?
- What if the alias or path is already in use by a View or used by a menu router?
- Is there a limit to the number of nodes that can be imported this way?
To create the export file, the node id is used to create the file. After that, the filename and 'unique id' references the alias of that node. So when you import the node, the node id on the production site will be determined by looking up the alias of the node. If a matching alias is found, that is the node that gets updated. If no matching alias is found, a new node gets created. The alias becomes the unique id.What are the risks of this import export model?
At present the known risks are:
- If the exported node uses entity references that do not exist on prod, the entity reference will either not be made, or reference an entity that is using that entity id on prod. This can be mitigated by exporting your source node while using a recent copy of the production DB.
- If the exported node uses taxonomy terms that do not exist on prod, the tag may import incorrectly. This can be mitigated by exporting your source node while using a recent copy of the production DB.
- if you are using pathato and the existing pattern on the production site is different than the pattern on your sandbox. The imported node will end up with a different alias, resulting in an invalid import. The imported node will be deleted since it failed validation and the hook_update_N will fail. This can be mitigated by exporting your source node while using a recent copy of the production DB.
- File attachments. There is currently not a way to bring attached files along with them unless the files already exist with a matching fid on production.
See answers 1 and 2 in What are the risks of this import export model?Does the import show up as a revision?
Yes it does, and the revison note contains the imported note, but also indicates it was imported with Hook Update Deploy Tools. The revision will take on the status of the exported node. If the exported node was unpublished, the impoirted revision will be unpublished.What happens if the import does not validate?
If the import was to an existing node, the update revision wil be deleted and return the node to its last published revision. If the import was for a node that did not exist on the site, the node and its first revision will be deleted. In either case, if the import was run through a hook_update_N, that update will fail and allow it to be re-run once the issue is resolved.What if the alias or path is already in use by another node?
If the alias is in use by a node, that node will be updated by the import. The alias is the unique id that links them not the nid.What if the alias or path is already in use by a View or used by a menu router?
If the alias is in use on the site by something other than a node, the import will be prevented. If the import is being run by a hook_update_N() then the update will fail and can be run when the issue is resolved.Is there a limit to the number of nodes that can be imported this way?
Technically, there is no real limit. Realistically, it is not a great workflow to move all of your content this way. It is not a good workflow. This export import method is best reserved for mission critical pages like forms or thankyou pages that go along with a Feature deployment. It is also good for pages that often get destroyed during early site development like style guides and example pages.